After the meteoric rise of generative artificial intelligence (GAI) models that transmute instructions into images, the new advance in this field are systems that generate clips from text. What Sora is like and what other tools are available for creating cinematic content.
From a technical point of view, according to OpenAI, what distinguishes Sora – which derives from the Japanese word “sky” – is the quality of its frames and also the ease of composing clips of up to a minute, longer than the fragments they offer. rivals.
Another amazing feature is that it doesn’t just produce short films that satisfy audience demands requiresbut does so in a way that demonstrates an emerging understanding of audiovisual narrative.
In turn, it includes a facial expression module, which allows you to animate figures with a high degree of emotion and gestures based on the input text, thus increasing the coherence of the images.
Tip: “Trailer for a film about the adventures of a 30-year-old astronaut wearing a red wool knitted motorcycle helmet, blue sky, salt desert, cinematic style, shot on 35mm film, bright colors”. pic.twitter.com/0JzpwPUGPB
—OpenAI (@OpenAI) February 15, 2024
And although they are not required previous knowledge In handling cameras or directing, Sora poses no threat to Hollywood, as films cannot be made by cooking 60-second snippets. The disruption will happen on TikTok and Instagram, where the brief predominates.
The other side of the coin is its simplicity in becoming a quick and effective solution spread misinformationmaking it even more difficult to distinguish the real from the illusory.
The previews that the company has shared on its website indicate that a new creative frontier could spread in this area. The biological texture of a squirrel’s fur, the fear reflected in an astronaut’s gaze or the irregular step of a woman wandering through the crowd are some examples of his detailed setting.
Through X, the tech brand explained that it is located in a research phase. They are working with members of what is known as Red Teamcomposed of academic experts in the field of security, to address content that incites entertainment and glorifies some prejudices.
And although Sora is the prototype that generates the most expectations – due to its relationship with GPT and Dall-e – there are other models already operational, such as Stable Video Diffusion, Runway, Wonder Studio, which are a few meters ahead.
Stable video streaming
Today we’re adding Stable Video Diffusion, our foundational model for generative video, to the Stability AI Developer Platform API.
The model can generate 2 seconds of video, comprising 25 generated frames and 24 FILM interpolation frames, in an average time of 41… pic.twitter.com/CSUh3BoZ1a
— AI Stability (@AIStability) December 20, 2023
This AI engine designed to compose images from text has extended its roots to videos. It belongs to the company Stability AI and its distinctive feature is being of Open Sourceso that any developer can review its structure, make changes and create tools from the original.
The model runs through deep neural networks, trained with thousands of hours of video. From a textual description or an image it is possible to create sequences of 14 to 25 frames, with customizable speeds from 3 to 30 frames per second and a maximum duration of 5 seconds.
One of its technical possibilities is to simulate the effect of a camera moving around an object or person. The same trick is used to approximate a face. In both cases it is also possible to adjust the speed at which the lens moves.
And although all these solutions are in a phase of full evolution, with permanent innovations carrying a greater risk in the results, the stitches are still visible in this application.
However, since it is a product that can be used free For non-commercial purposes, this unique feature gives it an extra level of indulgence that others don’t have.
“We are working daily to optimize the model with new versions that improve video performance and stability,” emphasizes Tom Mason, CTO of Stability.
runway
Generate anything. https://t.co/ekldoIshdw pic.twitter.com/o70Ls39E6R
— Pista (@pistaml) December 13, 2023
It is the most established platform, as its Gen 2 technology (soon to be replaced by Gen 3) has managed to overcome the methodological inaccuracies faced by many of its competitors.
One of its attractions is intervening on an already recorded video. That is, if you have a shot of a man walking his dog, you can ask him to replace the animal with a lion or replace the urban background with a jungle background.
Another option is to animate a video without providing too many instructions. Once loaded, the software applies its alchemy, and once returned, you can be asked to apply new changes or undo what you’ve done.
Runway ML includes the ability to generate 3D models from images and descriptions. This allows designers to materialize their three-dimensional ideas without the need for modeling knowledge.
There is a plan basic and free which limits video length to 16 seconds and includes 3 projects per month. The standard subscription costs $12 and the Pro subscription goes up to $28. Outside of this menu, the cost to adopt Gen-2 is 0.05 cents per second, which translates to $3 per minute.
Study of Wonders
We are proud to present Wonder Studio.
An AI tool that automatically animates, lights and composites CG characters in a live scene. No complicated 3D software, no expensive production hardware – all you need is a camera.
Sign up for the closed beta at https://t.co/QPyOVMjEAy pic.twitter.com/aHf2bLTsgZ
— Wonder Dynamics (@WonderDynamics) March 9, 2023
With a highly professional edge, the mission of this model suitable for independent productions is to simplify the film creation process through computer graphics (CGI), using artificial intelligence to make the process less cumbersome and more accessible.
Its developers claim that CGI no longer involves expensive motion capture equipment and that adding special effects It’s no more complex than applying filters in Photoshop.
With this platform you can easily produce films by selecting actors and assigning filler characters so that the system takes care of animating them and introducing them into each scene.
The advantage of Wonder Studio is its ability to resolve all the nuances of the production, from lighting, composition and camera movement, to the facial expression of the actors. With a single camera it is possible to obtain a complete and quality production.
To do this, it uses machine learning algorithms that draw a realistic 3D model of facial and body movements, allowing them to be coupled without being noticed in a video sequence.
Source: Clarin
Linda Price is a tech expert at News Rebeat. With a deep understanding of the latest developments in the world of technology and a passion for innovation, Linda provides insightful and informative coverage of the cutting-edge advancements shaping our world.