ElevenLabs, a voice AI startup co-founded by former Google machine learning engineer Piotr Dabkowski and ex-Palantir deployment strategist Mati Staniszewski, launches an AI model for generating detailed sound effects using text prompts.
This innovative leap forward fills the silent void left by OpenAI’s Sora, a state-of-the-art text-to-video model praised for its high-resolution video clips devoid of sound.
We were blown away by the Sora announcement but felt it needed something…
What if you could describe a sound and generate it with AI? pic.twitter.com/HcUxQ7Wndg
— ElevenLabs (@elevenlabsio) February 18, 2024
Following the buzz around Sora’s unveiling, which showcased the model’s capacity to create visually captivating videos from textual descriptions, ElevenLabs announced its pioneering project on February 18, 2024.
The project aims to transcend the limitations of current AI video technology by introducing an auditory dimension to the silent films produced by Sora.
Eleven Labs wrote in a recent blog post, “We used text prompts like “waves crashing,” “metal clanging,” “birds chirping,” and “racing car engine” to generate audio that we overlaid onto some of our favorite clips from the OpenAI Sora announcement. “We’re thrilled by the excitement and support from the community and can’t wait to get it into your hands,” it added.
By utilizing text prompts such as “waves crashing,” “metal clanging,” “birds chirping,” and “racing car engine,” ElevenLabs successfully launches realistic AI sound effects onto video clips, thus enriching the sensory experience of AI-generated content.
ElevenLabs, valued at over $1 billion following a successful $80 million Series B funding round, has been trying to eliminate linguistic barriers in content through its AI voice technologies.