The better question may be, what are you waiting for? 😉 Let’s roll, ChatGPT:
Fun! Here’s my guy Seamus as a sleepy dude. pic.twitter.com/JZ7XOT7Kws
— John Nack (@jnack) April 9, 2025
The better question may be, what are you waiting for? 😉 Let’s roll, ChatGPT:
Fun! Here’s my guy Seamus as a sleepy dude. pic.twitter.com/JZ7XOT7Kws
— John Nack (@jnack) April 9, 2025
Check out this fun toy:
[1/8] Drawing → 3D render with Gemini 2.0 image generation… by @dev_valladares + me
Make your own at the link belowhttps://t.co/sy8poJZYuQ pic.twitter.com/DkGT6DRUsb
— Trudy Painter (@trudypainter) April 4, 2025
Apparently I’m over my quota, so sadly the world will never get to see a Ghiblified rendering of my crudely drawn goldendoodle!

The team showed of good new stuff, including—OMG—showing how to use Photoshop! (On an extremely personal level, “This is what it’s like when worlds colliiiide!!”)
As it marks its 50th anniversary, Microsoft is updating Copilot with a host of new features that bring it in line with other AI systems like ChatGPT or Claude. We got a look at them during the tech giant’s 50th anniversary event today, including new search capabilities, Copilot Vision which will be able to analyze real-time video from a mobile camera. Copilot will also now be able to use the web on your behalf. Here’s everything you missed.
2025 marks an unheard-of 40th year in Adobe creative director Russell Brown’s remarkable tenure at the company. I remember first encountering him via the Out Of Office message marking his 15-year (!) sabbatical (off to Burning Man with Rick Smolan, if I recall correctly). If it weren’t for Russell’s last-minute intervention back in 2002, when I was living out my last hours before being laid off from Adobe (interviewing at Microsoft, lol), I’d never have had the career I did, and you wouldn’t be reading this now.
In any event, early in the pandemic Russell kept himself busy & entertained by taking a wild series of self portraits. Having done some 3D printing with him (the output of which still forms my Twitter avatar!), I thought, “Hmm, what would those personas look like as plastic action figures? Let’s see what ChatGPT thinks.” And voila, here they are.
Click through the tweet below if you’re curious about the making-of process (e.g. the app starting to render him very faithfully, then freaking out midway through & insisting on delivering a more stylized, less specific rendition). But forget that—how insane is it that any of this is possible??
Can you show ChatGPT 5 portraits of legendary Adobe creative director Russell Brown and get a whole set of action figures? Yep! pic.twitter.com/gLTIcGqLJ0
— John Nack (@jnack) April 4, 2025

It’s pretty stunning what a single creator can now create in a matter of days! Check out this sequence & accompanying explanation (click on the post) from Martin Gent:
I tried to make this title sequence six months ago, but the AI tools just weren’t up to it. Today it’s a different story. Sound on!
Since the launch of ChatGPT’s 4o image generator last week, I’ve been testing a new workflow to bring my characters – Riley Harper and her dog,… pic.twitter.com/SMgjDnJWH1
— Martin Gent (@martgent) April 3, 2025
Tools used:
People can talk all the smack they want about “AI slop”—and to be sure, there’s tons of soulless slop going around—but good luck convincing me that there’s no creativity in remixing visual idioms, and in reskinning the world in never-before-possible ways. We’re just now dipping a toe into this new ocean.
ChatGPT 4o’s new image gen is insane. Here’s what Severance would look like in 8 famous animation styles
1/8:
Rankin/Bass – That nostalgic stop-motion look like Rudolph the Red-Nosed Reindeer. Cozy and janky. pic.twitter.com/5rFL8SGttS— Bennett Waisbren (@BennettWaisbren) March 27, 2025
See the whole thread for a range of fun examples:
4/8:
Pixar – Clean, subtle facial animation, warm lighting, and impeccable shot composition. pic.twitter.com/FNWgPccHcI— Bennett Waisbren (@BennettWaisbren) March 27, 2025
It’s insane what a single creator—in this case David Blagojević—can do with AI tools; insane.
I’m blown away!
This KFC concept ad is 100% AI generated!
My friend David Blagojevic (he’s not on X) created this ad concept for KFC and it’s incredible!
Tools used: Runway, Pika, Kling AI, Google DeepMind Veo2, Luma AI, OpenAI Sora, upscaled with Topaz Labs and music… pic.twitter.com/u9ics8M51x
— Salma (@Salmaaboukarr) March 31, 2025
It’s worth noting that creative synthesis like this doesn’t “just happen,” much less in some way that replaces or devalues the human perspective & taste at the heart of the process: everything still hinges on having an artistic eye, a wealth of long-cultivated taste, and the willpower to make one’s vision real. It’s just that the distance between that vision & reality is now radically shorter than it’s ever been.
It’s funny to think of anyone & anything as being an “O.G.” in the generative space—but having been around for the last several years, Runway has as solid a claim as anyone. They’ve just dropped their Gen-4 model. Check out some amazing examples of character consistency & camera control:
Today we’re introducing Gen-4, our new series of state-of-the-art AI models for media generation and world consistency. Gen-4 is a significant step forward for fidelity, dynamic motion and controllability in generative media.
Gen-4 Image-to-Video is rolling out today to all paid… pic.twitter.com/VKnY5pWC8X
— Runway (@runwayml) March 31, 2025
Here’s just one of what I imagine will be a million impressive uses of the tech:
First test with @runwayml‘s Gen-4 early access!
First impressions: I am very impressed! 10 second generations, and this is the only model that could do falling backwards off a cliff. Love it! pic.twitter.com/GZS1B7Wpq0
— Christopher Fryant (@cfryant) March 31, 2025
Meanwhile Higgsfield (of which I hadn’t heard before now) promises “AI video with swagger.” (Note: reel contains occasionally gory edgelord imagery.)
Now, AI video doesn’t have to feel lifeless.
This is Higgsfield AI: cinematic shots with bullet time, super dollies and robo arms — all from a single image.
It’s AI video with swagger.
Built for creators who move culture, not just pixels. pic.twitter.com/dJdQ978Jqd
— Higgsfield AI (@higgsfield_ai) March 31, 2025
It’s so good, it’s bad! 😀
Currently asking ChatGPT for faux-German words like
Überintelligenzchatbotrichtigkeitsahnungsscham:“The bizarre cocktail of joy, panic, and existential dread a product manager experiences when an AI answers a tough product question better than they could.” pic.twitter.com/yVZVdoKZi9
— John Nack (@jnack) March 30, 2025
Seeing this, I truly hope that Adobe isn’t as missing in action as they seem to be; fingers crossed.
In the meantime, simply uploading a pair of images & a simple prompt is more than enough to get some compelling results. See subsequent posts in the thread for details, including notes on some shortcomings I observed.
A quick test of ChatGPT virtual product photography, combining real shoes with a quick render from @krea_ai/@bfl_ml Flux:
“Please put these shoes into the image of the basketball court, held aloft in the foreground by a man’s hand.” pic.twitter.com/k1AhTdHFcs— John Nack (@jnack) March 28, 2025
See also (one of a million tests being done in parallel, I’m sure):
Still experimenting with chatgpt4o
prompt: “model wearing cap provided”
not bad pic.twitter.com/FObSXeyxOS
— Salma (@Salmaaboukarr) March 26, 2025
We’re speed-running our way through the novelty->saturation->nausea cycle of Studio Ghibli-style meme creation, but I find this idea fresher: turn Ghibli characters into Dorothea Lange-style photos:
— Sterling Crispin (@sterlingcrispin) March 26, 2025
In the first three workdays of this week, we saw three new text-to-image models arrive! And now that it’s Thursday, I’m like, “WTF, no new Flux/Runway/etc.?” 🙂
For the last half-year or so, Ideogram has been my go-to model (see some of my more interesting creations), so I’m naturally delighted to see them moving things forward with the new 3.0 model:
I don’t yet quite understand the details of how their style-reference feature will work, but I’m excited to dig in.
Meanwhile, here’s a thread of some really impressive initial creations from the community:
We launched Ideogram 3.0 just three hours ago, and we’ve already seen an incredible wave of striking images. Here are 16 of our favorites so far:
1/ @krampus76 pic.twitter.com/tbwfMfkvg5
— Ideogram (@ideogram_ai) March 26, 2025
The family that bricks together, sticks together? 🙂
Worked great on this group shot as well—with the exception of disappearing one cousin! pic.twitter.com/pZzXfPurv3
— John Nack (@jnack) March 26, 2025
“Dress Your Family in Corduroy and Denim” — David Sedaris
“Turn your fam into Minecraft & GTA” — Bilawal Sidhu
Entire ComfyUI workflows just became a text prompt.
Open an image in GPT-4o and type “turn us into Roblox / GTA-3 /Minecraft / Studio Ghibli characters” pic.twitter.com/rCXclZklq5
— Bilawal Sidhu (@bilawalsidhu) March 26, 2025
And meanwhile, on the server side:
ChatGPT when another Studio Ghibli request comes in pic.twitter.com/NF5sy24GlU
— Justine Moore (@venturetwins) March 26, 2025
Nearly twenty years ago (!), I wrote here about how The Killing’s Gotta Stop—ironically, perhaps, about then-new Microsoft apps competing with Adobe. I rejected false, zero-sum framing then, and I reject it now.
Having said that, my buddy Bilawal’s provocative framing in this video gets at something important: if Adobe doesn’t get on its game, actually delivering the conversational editing capabilities we publicly previewed 2+ years ago, things are gonna get bad. I’m reminded of the axiom that “AI will not replace you, but someone using AI just might.” The same goes for venerable old Photoshop competing against AI-infused & AI-first tools.
In any case, if you’re interested in the current state of the art around conversational editing (due to be different within weeks, of course!), I think you’ll enjoy this deep dive into what is—and isn’t—possible via Gemini:
Specific topic sections, if you want to jump right to ’em:
The old (hah! but it seems that way) gal turns two today.
The ride has been… interesting, hasn’t it? I remain eager to see what all the smart folks at Adobe have been cooking up. As a user of Photoshop et al. for the last 30+ years, I selfishly hope it’s great!
Welcome to the world, #AdobeFirefly! https://t.co/R92lBktZIQ
We have great stuff you can out try right now, plus so much brewing in the lab. Here’s a quick preview: pic.twitter.com/hIaW9EpMor
— John Nack (@jnack) March 21, 2023
In the meantime, I’ll admit that watching the video above—which I wrote & then made with the help of Davis Brown (son of Russell)—makes me kinda blue. Everything it depicts was based on real code we had working at the time. (I insisted that we not show anything that we didn’t think we could have shipping within three months’ time.) How much of that has ever gotten into users’ hands?
Yeah.
But as I say, I’m hoping and rooting for the best. My loyalty has never been to Adobe or to any other made-up entity, but rather to the spirit & practice of human creativity. Always will be, until they drag me off this rock. Rock the F on.
Man, I’m old enough to remember writing a doc called “Yes, And…” immediately upon the launch of DALL•E in 2022, arguing that of course Adobe should develop its own generative models and of course it should also offer customers a choice of great third-party models—because of course no single model would be the best for every user in every situation.
And I’m old enough to remember being derided for just not Getting It™ about how selling per-use access to Firefly was going to be a goldmine, so of course we wouldn’t offer users a choice. ¯\_(ツ)_/¯
Oh well. Here we are, exactly two years after the launch of Firefly, and Adobe is going to offer access to third-party models. So… yay!
Even more news today! We are expanding our footprint in the @Adobe ecosystem to offer more choice to their creators pic.twitter.com/A4tHRkb25h
— Black Forest Labs (@bfl_ml) March 19, 2025
I guarantee you, the TTP for the feature is less than the length of this 45-second promo. :-p
Here’s a little holiday-appropriate experiment featuring a shot of my dad & me (in Lego form, naturally) at my grandmother’s family farm in County Mayo. Sláinte!
A little St. Paddy’s fun testing Google @GeminiApp‘s conversational editing abilities on Lego pics from Ireland: pic.twitter.com/LPCD0D3igi
— John Nack (@jnack) March 17, 2025
Speaking of reskinning imagery (see last several posts), check out what’s now possible via Google’s Gemini model, below. I’ve been putting it to the test & will share results shortly.
Alright, Google really killed it here.
You can easily swap your garment just by uploading the pieces to Gemini Flash 2.0 and telling it what to do. pic.twitter.com/pNPBkIdRqy
— Halim Alrasihi (@HalimAlrasihi) March 14, 2025
This enhanced capability, which apparently now uses a cloud-hosted model, looks really promising. See before & after:
The Photoshop Beta also has some pretty wild improvements to Remove Background pic.twitter.com/yu7u8ISbMW
— Howard Pinsky (@Pinsky) March 13, 2025
Another example:
https://t.co/VuXQVHMkN1 pic.twitter.com/mcy0nQ3b6m
— Howard Pinsky (@Pinsky) March 14, 2025
Another day, another set of amazing reinterpretations of reality. Take it away Nathan…
3 tests of Runway’s first frame feature. It’s very impressive and temporally coherent. Input is a video and stylized first frame. ✨
First example here is a city aerial to: circuit board, frost, fire, Swiss cheese, Tokyo. #aivideo #VFX pic.twitter.com/Y7HST74uBy
— Nathan Shipley (@CitizenPlain) March 6, 2025
…and Bilawal:
Playing guitar, reskinned with Runway’s restyle feature — pretty epic for digital character replacement.
I’m genuinely impressed by how well the fretting & strumming hands hold up.
Not perfect yet, but pulling this off would basically be impossible with Viggle or even Wonder… pic.twitter.com/UJBS9c8U1a
— Bilawal Sidhu (@bilawalsidhu) March 7, 2025
This temporally coherent inpainting is utterly bonkers. It’s just the latest—and perhaps the most promising—in myriad virtual try-on techniques I’ve seen & written about over the years.
This is effortless fashion
Made with @pika_labs Pikaswaps feature pic.twitter.com/BE9LDP8eAR
— Jessie_Ma (@ytjessie_) March 12, 2025
I love seeing the Magnific team’s continued rapid march in delivering identity-preserving reskinning
IT’S FINALLY HERE!
Mystic Structure Reference!
Generate any image controlling structural integrity Infinite use cases! Films, 3D, video games, art, interiors, architecture… From cartoon to real, the opposite, or ANYTHING in between!
Details & 12 tutorials pic.twitter.com/brw4Dx39gz
— Javi Lopez (@javilopen) February 27, 2025
This example makes me wish my boys were, just for a moment, 10 years younger and still up for this kind of father/son play. 🙂
Storyboarding? No clue! But with some toy blocks, my daughter’s wild imagination, and a little help from Magnific Structure Reference, we built a castle attacked by dragons. Her idea coming to life powered up with AI magic.
Just a normal Saturday Morning.
Behold, my daughter’s… pic.twitter.com/52tDZokmIT— Jesus Plaza (@JesusPlazaX) March 8, 2025
“Rather than removing them from the process, it actually allowed [the artists] to do a lot more—so a small team can dream a lot bigger.”
Paul Trillo’s been killing it for years (see innumerable previous posts), and now he’s given a peek into how his team has been pushing 2D & 3D forward with the help of custom-trained generative AI:”
Traditional 2d animation meets the bleeding edge of experimental techniques. This is a behind the scenes look at how we at Asteria brought the old and the new together in this throwback animation “A Love Letter to Los Angeles” and collaboration with music artist Cuco and visual… pic.twitter.com/3eWSdgckXn
— Paul Trillo (@paultrillo) March 7, 2025
Here’s a fun use of Flux->Minimax (see workflow details):
Fast food, but make it Lego.
byu/Sad-Ambassador-9040 incomfyui
A passing YouTube vid made me wonder about the relative strengths of World War II-era bombers, and ChatGPT quickly obliged by making me a great little summary, including a useful table. I figured, however, that it would totally fail at making me a useful infographic from the data—and that it did!
Just for the lulz, I then ran the prompt (“An infographic comparing the Avro Lancaster, Boeing B-17, and Consolidated B-24 Liberator bombers”) through a variety of apps (Ideogram, Flux, Midjourney, and even ol’ Firefly), creating a rogue’s gallery of gibberish & Franken-planes. Check ’em out.
Currently amusing myself with how charmingly bad every AI image generator is at making infographics—each uniquely bizarre! pic.twitter.com/U3cs8ySoVa
— John Nack (@jnack) March 6, 2025
Check out this delightful demo:
By combining @pika_labs Pikaframes and @freepik, I now have the magical ability to jump through space and time and in this example, music becomes a transformative element teleporting this woman to a new location. This is how it’s done. 1/6
The videos below are fully narrated… pic.twitter.com/06WtgI50ZV
— Travis Davids (@MrDavids1) March 3, 2025
Individual steps, as I understand them:
Another day, another ~infinite canvas for ideation & synthesis. This time, somewhat to my surprise, the surface comes from VSCO—a company whose users I’d have expected to be precious & doctrinaire in their opposition to any kind of AI-powered image generation. But who knows, “you can just do things.” ¯\_(ツ)_/¯
The British Academy Film Awards have jumped into a whole new dimension to commemorate the winners of this year’s awards:
The capturing work was led by Harry Nelder and Amity Studio. Nelder used his 16-camera rig to capture the recent winners. The reconstruction software was a combination of a cloud-based platform created by Nelder, which is expected to be released later this year, along with Postshot. Nelder further utilized the Radiance Field method known as Gaussian Splatting for the reconstruction. A compilation video of all the captures, recently posted by BAFTA, was edited by Amity Studio
[Via Dan Goldman]
Introducing FLORA, Your Intelligent Canvas.
Every creative AI tool, thoughtfully connected. pic.twitter.com/SUHrHtrQmn
— weber (@weberwongwong) February 26, 2025
Their pitch:
Behold the majesty (? :-)) of CapCut’s new “Microwave” filter (whose name makes more sense if you listen with sound on):
As I asked Bilawal, who posted the compilation, “What is this, and how can I know less about it?”
Slightly funky UI (I’d never have figured this out on my own), but amazing identity preservation! (Why can’t I do anything like this in Photoshop…?)
Check it out (probably easier to grok by watching vs. reading a description):
From the static camera feed, EditIQ initially generates multiple virtual feeds, emulating a team of cameramen. These virtual camera shots termed rushes are subsequently assembled using an automated editing algorithm, whose objective is to present the viewer with the most vivid scene content.
Tired: Random “slot machine”-style video generation
Inspired: Placing & moving simple guidance objects to control results:
Check out VideoNoiseWarp:
Every now and then something comes along that feels like it could change everything… NoiseWarp + CogVideoX lets you animate live action scenes with rough mockups!
ComfyUI nodes by @Kijaidesign https://t.co/AziU049jbg pic.twitter.com/eZsXJ38lxv
— Ingi Erlingsson (@ingi_erlingsson) January 21, 2025
Check out this fun mixed-media romp, commissioned by Adobe:
This video combines AI-generated elements (balloon, kite, surfboard, and backgrounds) with my own real-world practical effects and stop motion.
I made this for @Adobe Firefly and I’ll share tutorial tomorrow!
Thanks @Adobe for sponsoring my art #AdobePartner #AdobeFirefly pic.twitter.com/dPLrzCchH9
— Karen X. Cheng (@karenxcheng) February 12, 2025
And here’s a look behind the scenes:
Here’s the tutorial! This video combines AI-generated elements (balloon, kite, surfboard, and backgrounds) with my own real-world practical effects and stop motion.
I made this for #AdobeFirefly
Thanks @Adobe for sponsoring my art #AdobePartner pic.twitter.com/yUZtMlwk2r— Karen X. Cheng (@karenxcheng) February 13, 2025
The YouTube mobile app can now tap into Google’s Veo model to generate video, as shown below. Hmm—this feels pretty niche at the moment, but it may suggest the shape of things to come (ubiquitous media synthesis, anywhere & anytime it’s wanted).
For the longest time, Firefly users’ #1 request was to use images to guide composition of new images. Now that Firefly Video has arrived, you can use a reference image to guide the creation of video. Here’s a slick little demo from Paul Trani:
Firefly Video (beta) is now available to everyone! Give it a whirl and share your results!https://t.co/sOeN1pwXcV #adobefirefly #communityxadobe pic.twitter.com/ZOvkqKSq9T
— Paul Trani (@paultrani) February 12, 2025
These changes, reported by Forbes, sound like reasonable steps in the right direction:
Starting now, Google will be adding invisible watermarks to images that have been edited on a Pixel using Magic Editor’s Reimagine feature that lets users change any element in an image by issuing text prompts.
The new information will show up in the AI Info section that appears when swiping up on an image in Google Photos.
The feature should make it easier for users to distinguish real photos from AI-powered manipulations, which will be especially useful as Reimagined photos continue to become more realistic.
Conversational creation & iteration is such a promising pattern, as shown through people making ChatGPT take images to greater & greater extremes:
— No Context Shitposting (@NoContextCrap) February 8, 2025
But how do we go from ironic laughs to actual usefulness? Krea is taking a swing by integrating (I think) the Flux imaging model with the DeepSeek LLM:
Krea Chat is here.
a brand new way of creating images and videos with AI.
open beta out now. pic.twitter.com/dbHX31l92A
— KREA AI (@krea_ai) February 7, 2025
It doesn’t yet offer the kind of localized refinements people want (e.g. “show me a dog on the beach,” then “put a hat on the dog” and don’t change anything outside the hat area). Even so, it’s great to be able to create an image, add a photo reference to refine it, and then create a video. Here’s my cute, if not exactly accurate, first attempt. 🙂
Wow—check out this genuinely amazing demo from my old friend (and former Illustrator PM) Mordy:
In this video, I show how you can use Gemini in the free Google AI Studio as your own personal tutor to help you get your work done. After you watch me using it to learn how to take a sketch I made on paper to recreating a logo in Illustrator, I promise you’ll be running to do the same.
What the what?
this looks insane, MatAnyone
Stable Video Matting with Consistent Memory Propagation pic.twitter.com/tt1k23raYv
— AK (@_akhaliq) February 3, 2025
Per the paper,
We propose MatAnyone, a robust framework tailored for target-assigned video matting. Specifically, building on a memory-based paradigm, we introduce a consistent memory propagation module via region-adaptive memory fusion, which adaptively integrates memory from the previous frame. This ensures semantic stability in core regions while preserving fine-grained details along object boundaries.
I love it: nothing too fancy, nothing controversial, just a solid productivity boost:
Users can enter search terms like “a person skating with a lens flare” to find corresponding clips within their media library. Adobe says the media intelligence AI can automatically recognize “objects, locations, camera angles, and more,” alongside spoken words — providing there’s a transcript attached to the video. The feature doesn’t detect audio or identify specific people, but it can scrub through any metadata attached to video files, which allows it to fetch clips based on shoot dates, locations, and camera types. The media analysis runs on-device, so doesn’t require an internet connection, and Adobe reiterates that users’ video content isn’t used to train any AI models.
Goodbye, endless scrolling. Hello, AI-powered search panel. With the all-new Media Intelligence in #PremierePro (beta), the content of your clips is automatically recognized, including objects, locations, camera angles & more. Just input your search to find exactly what you need. pic.twitter.com/cOYXDKKaFI
— Adobe Video & Motion (@AdobeVideo) January 22, 2025
Check out this wild proof of concept from Trudy Painter at Google, and click into the thread for details.
Photos → Creative Code using Gemini
I built an experiment that turns photos into interactive @p5xjs sketches using Gemini 2.0 Flash.
Unlike UI generators, this creates code that mimics the *behavior* of what’s in the image – like smoke swirling or ripples spreading.
Check… pic.twitter.com/BbhYqUmZxA
— Trudy Painter (@trudypainter) January 23, 2025
Putting the proverbial chocolate in the peanut butter, those fast-moving kids at Krea have combined custom model training with 3D-guided image generation. Generation is amazingly fast, and the results are some combo of delightful & grotesque (aka “…The JNack Story”). Check it out:
God help you, though, if you import your photo & convert it to 3D for use with the realtime mode. (Who knew I was Cletus the Slack-Jawed Yokel?) pic.twitter.com/nuesUOZ1Db
— John Nack (@jnack) January 27, 2025
Check out this impressive demo that includes face swapping, selective editing, and conversion to video:
wow..
Freepik just launched AI Suite, it lets you edit images with FLUX PRO and..
you can even swap face to your own characters and animate them with top AI video generators.
step by step tutorial: pic.twitter.com/vLWHQAfHM1
— el.cine (@EHuanglu) January 25, 2025
Here’s another interesting snapshot of progress in our collective speedrun towards generative storytelling. It’s easy to pick on the shortcomings, but can you imagine what you’d say upon seeing this in, say, the olden times of 2023?
The creator writes,
Introducing The Heist – Directed by Jason Zada. Every shot of this film was done via text-to video with Google Veo 2. It took thousands of generations to get the final film, but I am absolutely blown away by the quality, the consistency, and adherence to the original prompt. When I described “gritty NYC in the 80s” it delivered in spades – CONSISTENTLY. While this is still not perfect, it is, hands down, the best video generation model out there, by a long shot. Additionally, it’s important to add that no VFX, no clean up, no color correction has been added. Everything is straight out of Veo 2. Google DeepMind
Here’s a nice write-up covering this paper. It’ll be interesting to dig into the details of how it compares to previous work (see category). [Update: The work comes in part from Adobe Research—I knew those names looked familiar :-)—so here’s hoping we see it in Photoshop & other tools soon.]
this is wild..
this new AI relighting tool can detect the light source in the 3D environment of your image and relight your character, the shadows look so realistic..
it’s especially helpful for AI images
10 examples: pic.twitter.com/sxNR39YTeT
— el.cine (@EHuanglu) January 18, 2025
Part 9,201 of me never getting over the fact we were working on stuff like this 2 years ago at Adobe (modulo the realtime aspect, which is rad) & couldn’t manage to ship it. It’ll be interesting to see whether the Krea guys (and/or others) pair this kind of interactive-quality rendering with a really high-quality pass, as NVIDIA demonstrated last week using Flux.
3D arrived to Krea.
this new feature lets you turn images into 3D objects and use them in our Real-time tool.
free for everyone. pic.twitter.com/b8gQMhUCN9
— KREA AI (@krea_ai) January 16, 2025
…featuring a dose of Microsoft Trellis!
Here’s how to create this cool 3D scene from a single image!
Midjourney (isometric image generation)
Trellis (Image to 3D Gaussian Splat)
Browser Lab (3D Editor Splat Import) pic.twitter.com/O1vJdaQRbc— IAN CURTIS (@XRarchitect) January 9, 2025
More about Trellis:
Powered by advanced AI, TRELLIS enables users to create high-quality, customizable 3D objects effortlessly using simple text or image prompts. This innovation promises to improve 3D design workflows, making it accessible to professionals and beginner alike. Here are some examples:
