EG "make me a 120 minute Star Trek / Star Wars crossover". It should be more or less comparable to a big-budget studio film, although it doesn't have to pass a full Turing Test as long as it's pretty good. The AI doesn't have to be available to the public, as long as it's confirmed to exist.
People are also trading
https://blog.google/technology/google-deepmind/ancestra-behind-the-scenes/
"To create “ANCESTRA”, Google DeepMind assembled a multidisciplinary creative team of animators, art directors, designers, writers, technologists and researchers who worked closely with more than 200 experts in traditional filmmaking and production, a live-action crew and cast, plus an editorial team, visual effects (VFX) artists, sound designers and music composers."
It's pretty cool that they did this, but it's a long way from generating a movie from a prompt. And the final result is only eight minutes long!
https://youtu.be/US2gO7UYEfY?si=vl5KCshO7cog4WUL
The WSJ made an AI video. But every shot required a lot of human effort to stitch outputs from different tools together and select a final clip that actually worked.
The script was also not AI generated, though I don't think that's the hardest part.
Total compute cost: ~$1000 for a three minute video.
@TimothyJohnson5c16 ...and the end result is way worse than a mediocre filmmaking student project. Can we have the market go to 10% already
@TimothyJohnson5c16 we're a month or two away from the halfway point of this market. I think this is a good marker of where things are now.
Sparkify https://sparkify.withgoogle.com/explore is another step in this direction
@ScottAlexander
What does "comparable to a big-budget movie" mean? It's possible for an AI to construct a compelling movie with strategic stylistic choices to overcome its limitations.
ie: A movie like Waking Life (2001) seems doable
Also to be extra pedantic - what makes a big budget movie? Like if I were to deepfake tom cruise into a few scenes would that automatically count?
Additionally, would it count if the AI took in additional data during the generation process
ie: The movie generates as people are are watching . While it generates, it simultaneously measures audience sentiment and updates the script as it is being made to calibrate on human preference.
As far as the UX setup flow goes, all it took was a prompt
Additionally+1: What is the domain breadth of possible generated movies for this to count?
ie: I could overfit an auto encoder to act as a a video player for Star Trek - where any prompt will output the same full high quality movie.
Additionally+2: How distinct does the movie have to be from existing media?
ie: It's possible to make a system that maps any input prompt and outputs the most similar existing high budget movie.
Additionally+3: How well does the movie have to adhere to the input prompt.
@TheAllMemeingEye I'm just not impressed by 10 different robots, 10 different kangaroos, the voices change, environments change.
Progress, but so much more needs to be done.
@TheAllMemeingEye I'm buying no in response to the market going up. I have no idea about the current state of progress, but trading the wiggles has rarely failed me in the past.
But I would expect a few bursts of seemingly impressive progress, even if the market ultimately resolves NO.
A lot of people have posted examples but this is really spooky good
https://x.com/simone_m_romeo/status/1926017107851960453?t=k_kxuTMAErVn4S1FFZfJHQ&s=19