- The AI Upload
- Posts
- 👾 AI Commercials, MCP Servers & More
👾 AI Commercials, MCP Servers & More
Everything you need to know about AI this week.


/// System Override
A tool, shift, or insight that challenges the default and demands a deeper look. Our main story of the week.
Welcome to the first edition of The AI Upload!
For this edition, I wanted to dive into a post I made recently on my socials that received an incredible response.
After playing with the new image generation capabilities of ChatGPT 4o, I wanted to see how far I could push it.
In collaboration with a brand that reached out, Mouthology, I created the follow “commercial” using nothing but AI.
🤯 I made this full commercial using only AI.
No film crew. Just 4 pictures, 4o, Kling, and a little editing.
Here’s how I pulled it off—and how you can too 👇
Bonus: Comment 'Prompt' and I'll DM you my private CustomGPT I used to produce the source images.
— Cory Dobbin (@CoryDobbin)
7:17 PM • Apr 7, 2025
So, here’s the sauce — exactly how I made it.
Phase One - Prep Work
To produce this ‘commercial’, I had to start with assets from the brand. I had Mouthology’s founder send me over images from a recent photoshoot, which ended up being really important for two reasons:
Having high resolution images to work with tend to yield better results with 4o’s image gen, in my experience, especially for product insertion
Having models holding the product helped me imagine a flow for the video - having the model start here, then move there, and replicate the face through scenes. Note: there were actually 2 different women in the photoshoot, but the end result has 3. The AI took over and added an extra model to the shoot. I’m not complaining.
These are the images I started with:
![]() | ![]() | ![]() |
The workflow I used for this was:
Input Images
CustomGPT Prompt
4o Image Gen
Animation Start & End Image
Kling AI For Animation or Higgsfield For Animation
Premiere Pro For Post Editing
So, to prepare, I needed to create images to use in Kling AI. One thing that ended up being very useful was Kling AI’s feature that allows you to set a Start and End point for your animation, so I wanted to create these images with 4o first.
Phase Two - Prompts & Base Images
I put the source images into a CustomGPT I built for producing better image prompts. I found that using this helped yield better results when prompting for a base image to use before animating it.
Here’s a link if you want to check it out for yourself: https://chatgpt.com/g/g-67eb10abb1ac8191882b6e6b5d18e33b-image-prompt-generator
Here’s an example of it turning a simple input into something easier for 4o to understand:

Once you produce the prompt using the CustomGPT, bring it to a new 4o chat, re-insert your source image along with the new prompt to get the base for animation. Using the prompt above, here’s what I received as the output. Exactly what I was hoping for.

Simply repeat this process for every scene you want in your ‘commercial’, then we can move on to the next section.
Phase Three - Animation
My tool of choice for this is Kling AI. Sometimes people ask why I pick this over Sora, and the answer is, unfortunately, not very sophisticated - Sora hasn’t been available for me to use for whatever reason, so Kling AI gets my money.
Higgsfield is another option, and has some very cool and unique offerings for cinematic camera movements, but I find it much less reliable than Kling, and takes much longer to render. In fact, the only scene in the commercial that used Higgsfield was the intro clip. I used the ‘FPV Done’ shot on the image above and instructed it to fly past the model into the clouds. I then reversed that in post to get the result I wanted - there’s an example in the thread on X linked above. Aside from this, I found Kling to be more reliable in producing the result I wanted.
As mentioned before, in Kling AI I animated using a start and end point so that I could give the commercial a feeling of continuity. For each clip, I would start it with the image that I used in the end of the previous clip so that when I put them into Premiere they would play through seamlessly.
The trick with Kling AI, in my experience, is to have the paid VIP version. When you subscribe, you get the ability to go ‘VIP’ when rendering, as well as a bunch of credits. It’s honestly quite affordable, and only cost ~$10 to produce all the clips I needed. When you run the clip as ‘VIP’ it costs 35 credits instead of 20 credits, so it’s significantly more expensive, but the result is better 9 times out of 10.
Aside from that, there’s no real hack here. You can give it a prompt to follow but I rarely found it stuck to the formula provided. Giving it a loose guideline seemed to work best, and keeping the clip short (5 seconds) allowed me to control the flow better. You can also adjust the ‘creativity’ level of the rendering to give it more flexibility, but I find this is delicate and best used only when you aren’t trying to retain detail in a product.
Once you’re done producing all of your clips, you’re ready for the final stage.
Phase Four - Post Processing
This part is both the easiest conceptually and complex and time consuming in execution. As you may have noticed, there’s a lot of movement in the commercial. This is because I created a speed ramp for each clip based on where it would fit best. This gives the movement more ‘glide’ and flow.
With all of the clips laid out in order in Premiere. I began to individually edit the speed ramp and duration of each clip, and lined them up. I created an Adjustment Layer to change the details of the clip (aesthetically, ie brightness, contrast, white point, etc). I also created the music using Splice by simply dragging in layers of the same tempo and key until I found the result I liked.
Finally, a few transition effects and that’s it! Like I said, complex in execution, but fundamentally simple in concept. Simple put: edit the clips together.

The end result was amazing, in my opinion - yes I am biased, but I don’t care, it’s sick. Given the amount of reach this post had, there were a few trolls that came out of the woodwork to share their thoughts on how this frame wasn’t perfect or that frame had an artifact, but nearly all of them missed the point.
This is what we can do with AI today. 4o is barely two weeks old. Give this workflow and tech stack another 6 months and it will be indistinguishable from reality. Getting your practice in now will make a world of difference in getting ahead for when that time comes.
If you use this workflow, let me know! I’d love to see what you cook up. Have fun out there!
To The Future,
Cory

/// Patch Notes
A log of new tools, major announcements, and news worth knowing, compressed for fast parsing.
ChatGPT Knows All
ChatGPT's memory feature is now available, and allows the AI to retain user-provided info across sessions for a more personalized experience. Users can manage what’s remembered, including editing or deleting details. The update moves ChatGPT closer to functioning like a true digital assistant, capable of maintaining context over time.
WhatsApp MCP Server Gets A Refresh
WhatsApp MCP server now supports sending and receiving images, videos, and voice notes, enhancing its multimedia capabilities.
The WhatsApp MCP server can now send and receive images, videos, and voice notes
Combine it with the new ElevenLabs MCP server to give it superpowers — using AI to transcribe the voice notes and send audio messages with 3,000+ voices
— Luke Harries (@LukeHarries_)
5:54 PM • Apr 7, 2025
Cursor Errors Meet Their Match
AI Jason outlines a method to reduce 90% of errors in AI coding environments like Cursor by implementing structured task management.
ElevenLabs Gets An MCP Server
ElevenLabs has launched its official MCP server, enabling AI tools like Claude and Cursor to access the full ElevenLabs AI audio platform through simple text prompts. This integration allows developers to seamlessly incorporate advanced voice synthesis into their applications.
Introducing the official ElevenLabs MCP server.
Give Claude and Cursor access to the entire ElevenLabs AI audio platform via simple text prompts.
You can even spin up voice agents to perform outbound calls for you — like ordering pizza.
— ElevenLabs (@elevenlabsio)
5:42 PM • Apr 7, 2025
Gemini Now Supports YouTube
Gemini now supports YouTube videos directly, enabling users to transcribe, translate, summarize, rewrite for tweets, and extract key hooks—all from a single link.
Gemini now supports YT videos directly.
- Transcribe and translate.
- Summarize and rewrite for a tweet.
- Extract the hooks.
- Describe the setting of the video.All you need to do is go to Gemini AI Studio and put the youtube link over there.
— Jean Gatt (@jean__gatt)
2:47 PM • Apr 7, 2025
Building Blocks Drop
The Claude MCP server now supports image inputs, allowing users to send images directly to Claude for analysis and response. This enhancement broadens the server's capabilities, enabling more interactive and versatile AI interactions.
Introducing blocks.so
A library of blocks/components that you can copy and paste into your apps.
Built on top of shadcn. Beautifully designed. Open Source.
— Duncan (@EphraimDuncan_)
2:40 PM • Apr 7, 2025

Thanks for reading this week’s edition of The AI Upload. Is there anything you want to see added? Something we missed? We want to hear your feedback! Simply reply to this email and we will read it. Thank you for your support!