New AI model by Meta elevates video editing

Movie Gen from Meta outperforms leading models in benchmarks, maintaining video structure and realism across diverse editing tasks.

Meta expands its AI toolkit with faster models and new datasets.

Meta has launched Movie Gen, a powerful AI model designed to produce 1080p videos with synchronised audio. The system can edit videos based on instructions, allowing for personalised content creation using user-supplied images.

With a transformer model containing 30 billion parameters, Movie Gen can generate 16-second videos at 16 frames per second. The model’s advanced techniques improve video motion realism, trained on over 100 million video-text pairs and 1 billion image-text pairs.

Movie Gen outperforms previous models, including Runway Gen3 and OpenAI Sora, particularly in video editing and text-to-video tasks. Benchmarks show its superiority in maintaining video structure and fine details, especially in the TGVE+ test.

Future developments for Movie Gen include improving scene understanding, safeguarding against misuse, and making the system more accessible. Meta envisions applications in social media, film production, and marketing campaigns.