- AI For Humans - The Newsletter
- Posts
- Kling 2.0 AI Video Gets Us Ever Closer to Prompt-To-Hollywood
Kling 2.0 AI Video Gets Us Ever Closer to Prompt-To-Hollywood
Also, Is OpenAI *really* working on a new social network for 4o Image Gen?
Today on AI For Humans The Newsletter!
Kling 2.0 is here! It’s good… but expensive
OpenAI Rumored Plan To Launch a Social Network…Plus, GPT-4.1
NVIDIA’s new plan to make AI chips in Texas
Plus, our can’t-miss AI feature of the week!
Welcome back to the AI For Humans Newsletter!
So… how far are we really away from a fully generated AI movie?
It’s a question that Kevin & I have been asking ourselves for a while. In the last few weeks, we’ve taken some significant steps forward (and probably cut a few months off our timeline) with the launches of Runway Gen-4, Google’s Veo 2, new tools from Higgsfield AI and now, Kling 2.0.
⚡️ Massive Update Just Dropped: Phase 2.0 for Kling AI!
🎥 KLING 2.0 Master for video generation, 🖼️ KOLORS 2.0 for image generation, 🎮 Multi-Elements Editor, 🎨 Image Editing & Restyle...
Kling AI 2.0 is all about empowering creators to bring meaningful stories to life — with— Kling AI (@Kling_ai)
7:08 AM • Apr 15, 2025
Kling is the best AI video model on the market right now, full stop.
I’ve been doing a bunch of video gens for <secret project> and kept coming back to Kling’s last model, Kling 1.6. It does the best job of giving lifelike animations to subjects and the sorts of subtle camera movements that can add a ton to your AI video productions.
What does 2.0 add to the mix?
Overall upgraded shot consistency, better physics (specifically lots of animals running), prompt coherence (specifically camera shots) & more. They’ve also launched a new update their AI Image model Kolors as well.
My first @Kling_ai 2.0 output was VERY good.
This image is particularly hard to use for video (I'll share the sora & runway outputs below) but for ten second single shot output?
it's really, really solid.
— AI For Humans Show (@AIForHumansShow)
3:43 PM • Apr 15, 2025
So if you’re an up-and-coming AI video filmmaker why wouldn’t you use Kling 2.0? Well, it’s expensive… 100 credits per 5 second output (up from 70 credits prior). Depending on what subscription level you’re on or if you’re buying credits, that’s a big price hike. But still, that’s significantly less than what Google is charging for VEO 2.

On the most expensive monthly plan, one 5-second gen will cost you $0.81
A slightly thornier issue: Kling is made by Kuaishou Technologies, a large Chinese tech firm, and if you have issues with how AI video models are trained then you may want to avoid it. The hard truth right now is that the Chinese video models (including Minimax) are some of the best and it’s more than likely these were trained on a massive corpus of films and TV shows. But it’s impossible to know because none of these companies (including the American ones) are opening their datasets for us to see.
We’re excited to dig in a bit more and push Kling to see what it does and, I guess, open our wallets to get a few more credits to put into the ol’ machine. Have I mentioned lately we have a Patreon tip jar for this exact reason? 😢
More on Thursday!
-Gavin (and Kevin)
3 Things To Know Today
Is OpenAI About To Launch a Social Network? Plus, GPT-4.1 Launches
Breaking news from The Verge’s Kylie Robison this AM: It sounds like OpenAI is thinking about a social network, specifically around images.
SCOOP: OpenAI is working on its own X-like social network, according to multiple sources familiar with the matter. While the project is still in early stages, we’re told there’s an internal prototype focused on ChatGPT’s image generation that has a social feed.
— Kylie Robison (@kyliebytes)
3:48 PM • Apr 15, 2025
We actually think this is a great idea to help discoverability of image prompts and to get creators seen. The Sora home page is a bit of a mess right now. Could this also be how they pay creators eventually?
Also, OpenAI is widely expected to launch their o3 model at some point this week and just in time, as Google Gemini 2.5 Pro has taken the AI lead across MANY use cases like coding & deep research. But before that, they decided to GO BACKWARDS (kind of) and are releasing a new version of GPT-4o called GPT-4.1 but only for developers.
Announcing GPT-4.1, GPT-4.1 mini, and GPT-4.1 nano in the API.
TL;DR: Major improvements on coding, instruction following, and long context. 💥
00:00 Intro
02:18 Coding
04:53 Instruction following
06:58 Long context
10:22 Demos, pricing, and availability
20:00 @windsurf_ai— OpenAI Developers (@OpenAIDevs)
7:11 PM • Apr 14, 2025
What is going on here? How many names can OpenAI really roll out? The basics: GPT-4.1 is significantly cheaper for developers, performs better than some reasoning models and in some cases, even beats GPT-4.5.
Google Created An AI To Talk To Dolphins
Have you ever wanted to talk to your dog? No, I’m not counting that one Tiktok dog who pushes the buttons. I mean really understand what they’re saying?
Well, Google DeepMind is hoping its work on DolphinGemma, a new AI designed specifically to understand what dolphins are saying, may get us to that dog-speaking universe.
Meet DolphinGemma, an AI helping us dive deeper into the world of dolphin communication. 🐬
— Google DeepMind (@GoogleDeepMind)
1:03 PM • Apr 14, 2025
NVIDIA’s Big Texas Move: AI Supercomputers Made in the USA
This has been quite the week in the world as your 401k can tell you. As markets are getting spooked about who’s going to place tariffs on whom, NVIDIA announced they're teaming up with Foxconn and Wistron to mass-produce AI supercomputers in Texas.
Why should you care? Well, bringing production stateside means more control, faster innovation, and fewer supply chain headaches. Is this the AI boom’s "made in America" moment? It might be the start of it.
We 💛 This - OpenAI’s 4o Image Gen Makes Pets into People
While OpenAI might be cracking down further on what we can and can’t do with 4o Image Gen (I was able to get this pretty fun vintage Barbie ad prompt to work), people are still discovering wild real world ways to push the technology.

This is my dog, Ollie. I also think he’s the second bassist in Phish.
Even though this prompt has gone fully mainstream (both of my daughters were doing it without any input on my part), it shows the insane power of what 4o Image Generation can accomplish. Not that long ago, you would’ve need a whole heck of a lot of weird ComfyUI spaghetti strings to pull something like this off and now, it’s as easy as asking ChatGPT. Just use the prompt in the image above and upload a photo of your pet.
Are you a creative or brand looking to go deeper with AI?
Join our community of collaborative creators on the AI4H Discord
Get exclusive access to all things AI4H on our Patreon
If you’re an org, consider booking Kevin & Gavin for your next event!