Greetings from Terminal 4 at JFK. I'm heading to Los Angeles this morning for an in-person speaking engagement tomorrow.
, a startup known for its AI voice cloning and text-to-speech software, is set to launch a new AI model that can generate realistic sound effects for videos based on text prompts. The company demonstrated this technology by adding background sounds to video clips created by OpenAI's Sora model (which, if you haven't seen yet, stop reading and ), showcasing the continued advancement of AI in creating immersive multimedia experiences.
The new AI model is expected to be able to create a variety of sounds – including footsteps, waves, and ambience – to accompany silent video footage. While the public release date has not been announced, interested individuals can about the launch.
As you can imagine, generative AI sound effects will impact audio and video production, gaming, and all forms of extended reality, where realistic sound effects play a crucial role in enhancing the overall user experience. The idea of data-driven hyper-personalized sound effects, alarms, and notifications has been talked about for years. Now it's just months away. This is super exciting!
As always your thoughts and comments are both welcome and encouraged. Just reply to this email. -s [email protected]
Mastering the Art of Prompt Engineering is not optional. Here's a to jump start your journey.
ABOUT SHELLY PALMER
Shelly Palmer is the Professor of Advanced Media in Residence at Syracuse University’s S.I. Newhouse School of Public Communications and CEO of The Palmer Group, a consulting practice that helps Fortune 500 companies with technology, media and marketing. Named he covers tech and business for , is a regular commentator on CNN and writes a popular . He's a , and the creator of the popular, free online course, . Follow or visit .