New AI model by Meta elevates video editing
Movie Gen from Meta outperforms leading models in benchmarks, maintaining video structure and realism across diverse editing tasks.
![Meta expands its AI toolkit with faster models and new datasets.](https://diplo-media.s3.eu-central-1.amazonaws.com/2023/09/meta-logo-metaverse-product-setting-podium-abstract-minimalistic-placement-abstract-background-1024x634.jpg)
Meta has launched Movie Gen, a powerful AI model designed to produce 1080p videos with synchronised audio. The system can edit videos based on instructions, allowing for personalised content creation using user-supplied images.
With a transformer model containing 30 billion parameters, Movie Gen can generate 16-second videos at 16 frames per second. The model’s advanced techniques improve video motion realism, trained on over 100 million video-text pairs and 1 billion image-text pairs.
Movie Gen outperforms previous models, including Runway Gen3 and OpenAI Sora, particularly in video editing and text-to-video tasks. Benchmarks show its superiority in maintaining video structure and fine details, especially in the TGVE+ test.
Future developments for Movie Gen include improving scene understanding, safeguarding against misuse, and making the system more accessible. Meta envisions applications in social media, film production, and marketing campaigns.