Tech
Google DeepMind unveils V2A, a new AI model that can generate soundtrack and dialogue for videos
Google DeepMind, the company’s AI research lab recently unveiled V2A, a new model that can generate audio from videos.
Video generation models like Sora, Dream Machine, Veo and Kling are advancing at a rapid pace, allowing users to generate videos from text prompts. But, the majority of these systems are limited to silent videos. Google DeepMind seems to be aware of the problem and is now working on a new large language model that can generate soundtracks and dialogues for videos.
In a blog post, the tech giant’s AI research lab unveiled V2A (Video-to-audio), a new work-in-progress AI model that “combines video pixels with natural language text prompts to generate rich soundscapes for the on-screen action.”
You have exhausted your
monthly limit of free stories.
Read more stories for free
with an Express account.
Invest in democracy. Full access to Express at just Rs 999/year
This premium article is free for now.
Register to read more free stories and access offers from partners.
Invest in democracy. Full access to Express at just Rs 999/year
This content is exclusive for our subscribers.
Subscribe now to get unlimited access to The Indian Express exclusive and premium stories.
Compatible with Veo, a text-to-video model the company introduced at the recently concluded Gooogle I/O 2024, V2A can be used to add dramatic music, realistic sound effects and dialogue that matches the tone of the video. Google says the new large language model also works with “traditional footage” like silent films and archival material.
The new V2A model can generate an “unlimited number of soundtracks” for any video and features an optional ‘positive prompt’ and ‘negative prompt’, which can be used to tune the output to your preferences. It also watermarks the generated audio with SynthID technology.
DeepMind’s V2A technology takes the description of a sound as input and uses a diffusion model trained on a combination of sounds, dialogue transcripts and videos. Since the model wasn’t trained on a lot of videos, the output can be distorted at times. Google also says it won’t release V2A to the public to prevent misuse anytime soon.
© IE Online Media Services Pvt Ltd
First uploaded on: 18-06-2024 at 17:10 IST