Audioshake, the leader of AI Sound Separation Technology, has launched multi-speaker.
The company says it is a powerful new model designed to separate an unlimited number of speakers into individual audio tracks.
This is the first model to achieve multi-speaker separation with high-resolution audio that opens up new creative use of audio AI, film, podcasts, UGC and TV content.
Wondercraft integrates Audioshake's multi-speaker into the audio studio, separating the generated podcasts from NotebookLM into individual speaker tracks, giving you more control over your conversations and final editing.
Multi-speaker leverages Audio Shark's unique AI technology to handle complex audio environments, including crowd dialogue, panel discussions, and fast-paced interviews. Separate them into individual speaker streams.
This model allows users to easily isolate individual speakers, improve transcription and caption accuracy, enable more accurate editing workflow, separate audio for audio AI tasks, and clean up duplicate dialogs for dubbing and localization.
“The launch of multi-speaker pushes the boundaries of what can be done with sound separation,” said Jessica Powell, CEO of Audioshake. “This model is designed for professionals working with complex audio mixes, such as broadcast, film or transcription.
Multi-speakers make it easier than ever to work with voices that were previously impossible to separate. ”
Audioshake's research director Fabian-Robert Stotter highlighted how the new model was designed to handle real-world scenarios. Our team worked to create solutions that are not only robust but accurate, even in a very challenging environment. ”
The multi-speaker model represents a major advancement for media and content industry professionals. By providing powerful tools to isolate overlapping voices, it enhances both workflow efficiency and audio clarity for applications, including:
Media and Entertainment: Even a chaotic soundscape achieves cleaner dialogue tracks and improves the overall listening experience for localization and dubbing of the audience: Translators and narrators can work with accurate, isolated audio tracks. Live broadcasts and events in summary: Broadcasting stations can extract clear voices for clear speeches during interviews, sports commentary, and panel discussions, improve audience engagement, and understand AI voice integration and research.
Multi-Speaker is now available through Audioshake's web-based platform and API. To contact us or experience it in person, please contact info@audioshake.ai.