š” Andromeda is now available on LALAL.AI desktop and mobile applications! (Stems: Vocal & Instrumental and Voice & Noise) And yes, it's accessible to everyone by default and included in your current plan. Regardless of the device you're running LALAL.AI on, you can now enjoy: ā Up to 40% faster processing, enabling more efficient workflows for creators and audio teams. ā 10% improvement in SDR, resulting in cleaner, more accurate separations. ā The frequency range remains consistent at up to 22 kHz with stereo processing, providing high-fidelity separation the industry expects. Andromeda handles challenging audio scenarios and complex instrument mixes with greater precision, reducing artifacts and increasing consistency ā Thanks to advanced DSP (Digital Signal Processing) techniques, Andromeda delivers consistent separation across quiet or loud tracks, letting creators and audio engineers work across different types of audio sources without worrying about volume inconsistencies. ā Reduced need for manual DAW cleanup, thanks to more precise layer separation and improved handling of complex mixes. ā No more trade-offs between detail and bleed control. Andromeda eliminates the former Clear Cut and Deep Extraction modes, giving you clean, detailed stems in a single pass. Just make sure your app's version is 2.11 š¤ And share your feedback on how the new algo works for you!
LALAL.AI
Software Development
AI-powered stem separation for professionals. Extract vocals, instruments, and dialogue with industry-leading precision.
About us
LALAL.AI is an advanced AI-powered stem separation service designed for music professionals, video editors, post-production studios, and content creators. Our technology enables businesses and creators to extract vocals, instruments, and dialogue with unmatched precision. ā Studio-grade accuracy powered by proprietary Perseus AI transformer models ā Next-gen stem separation: Isolate vocals, drums, bass, and 10+ stems with precision ā Voice cleaning & echo removal: Restore and enhance audio with a single click ā Trusted by 6,5M+ users across 150+ countries and major creative industries ā Used by content creators, editors, labels, video production industry leaders, and advertising pros for fast, clean post-production ā Integrates easily via API for B2B, enterprise, and SaaS platforms. ā SLA & 99.9% Uptime Guaranteed. GDPR-compliant We've compared all available audio splitting methods and proved that ours is the best. For the sake of objectivity, we have improved some of these methods using our music database ā the studio-quality multi-track recordings (the material that sound engineers operate with). We plan to continue experimenting with API integration for audio splitting and other purposes and share new ideas and solutions to help to make the life of millions of people easier. Want to integrate LALAL.AI? Explore our API to bring AI-powered audio separation to your platform.
- Website
-
https://www.lalal.ai
External link for LALAL.AI
- Industry
- Software Development
- Company size
- 11-50 employees
- Headquarters
- Zug
- Type
- Privately Held
- Founded
- 2020
- Specialties
- vocal extraction, audio production, sound production, karaoke, music production, vocal remover, AI, artificial intelligence, Music, audio splitter, machine learning, video, voice cleaner, video production, stem splitter, and audio
Locations
-
Primary
Get directions
Zug, CH
-
Get directions
Bahnhofstrasse 32
Bahnhofstrasse 32
Zug, Zug 6300, CH
Employees at LALAL.AI
Updates
-
Andromeda Removes a Long-Standing Trade-Off in Stem Separation Our newest, fastest, and clearest algorithm to date, Andromeda, removes the need to choose betweenĀ Clear CutĀ andĀ Deep Extraction, delivering detailed stems with minimal bleed in a single pass, fewer processing runs, less DAW cleanup, faster deliveries for teams and studios. Read more in the carousel below & try Andromeda on LALAL.AI to test its stem separation quality.
-
š§ Introducing Andromeda: LALAL.AIās Most Advanced Separation Model to Date & New Benchmark in Stem Separation Weāre pleased to announce the release of Andromeda, the latest evolution of our audio source separation technology that interprets audio tracks with near-human precision.Ā Built on six generations of research and extensive DSP (Digital Signal Processing) work, Andromeda takes a more nuanced approach, analyzing audio in terms of time, frequency, and tone, rather than simply processing sound as a single waveform. Key improvements that matter: ⢠Up to 40% faster processing, enabling more efficient workflows for creators and audio teams. ⢠10% improvement in SDR, resulting in cleaner, more accurate separations. ⢠The frequency range remains consistent at up to 22 kHz with stereo processing, providing high-fidelity separation the industry expects. ⢠Thanks to advanced DSP techniques, Andromeda delivers consistent separation across quiet or loud tracks. ⢠Reduced need for manual DAW cleanup, thanks to more precise layer separation and improved handling of complex mixes. ⢠No more trade-offs between detail and bleed control. Andromeda eliminates the former Clear Cut and Deep Extraction modes, giving users clean, detailed stems in a single pass. Andromeda is now available across key LALAL.AI tools, including Stem Splitter (Vocal & Instrumental and Voice & Noise stems), Lead & Back Vocal Splitter, Echo & Reverb Remover, and Voice Cleaner. The new algorithm is also available via API for Stem Splitter & Voice Cleaner clients. š Black Friday Special To mark the release, weāre offering a limited-time Black Friday Sale valid until December 5. Don't miss the chance to celebrate the release with us!
-
-
š”Do you know that not long ago, YouTubeĀ introduced automatic dubbing, a feature designed to help creators connect with audiences across different countries? If youāve come across videos dubbed this way, youāve probably noticed how good the translation is. At the same time, itās hard not to notice how unnatural or even robotic these dubbed voices often sound. Itās especially noticeable when youāre watching a dubbed video from a creator whose real voice youāve known for years before the auto-dubbing feature was introduced. The difference can feel jarring. The good news is that creators can now make a digital copy of their own voice and use it to dub their videos. The technology behind this is called voice cloning, and you, as a YouTube creator, video editor, or a dubbing team, can use it to dub YouTube videos in your own voice. Read how in our new article š
-
Honored to have been part of the incredible Music China 2025 in Shanghai! It was a true privilege to contribute to this unique platform where Eastern and Western music technologies converge, creating new possibilities for the future of sound. Our Head of China Market, Anna Lyashchenko, was thrilled to connect with brilliant minds and explore how LALAL.AI can help shape the next chapter of music creation. It was great to catch up with: ā Joya Hoya (禾ę„) from Positive Grid to discuss jamming technology; Xu Tao (å¾ę») fromĀ WeiKing Acoustics to talk about the specifics of real-time vocal music separation; Richard Gu from Replica to explore possibilities of the LALAL.AI VST plugin. We also connected with innovators from: Yooba Technology, iCON, ACE studio, Antelope Audio, å„ęµŖAPP, and A&C Pro Audio. The synergy from these meetings has given us incredible momentum to push the boundaries of audio AI even further. Catch you next time, Shanghai! ā ē“å°äøę¬”, äøęµ·ļ¼
-
-
LALAL.AI reposted this
Quando apro eventi di grandi dimensioni come DJ, mi piace spesso iniziare con un brano in versione orchestrale. Questa volta ho voluto sperimentare con lāintelligenza artificiale per creare una versione orchestrale di āMan I Needā di Olivia Dean (se cercate unāidea regalo per Natale, questo album ĆØ una bella scoperta). **IL PROCESSO CHE HO SEGUITO:** Ho iniziato usando Claude per elaborare un documento di 47 pagine contenente tutte le istruzioni tecniche di Suno (lāAI generativa per contenuti musicali). Questo mi ha permesso di creare le basi per un prompt ottimale. Successivamente, ho creato un progetto su ChatGPT con queste istruzioni come fondamento. La richiesta era semplice: ārealizza un prompt per Suno di una versione orchestrale con un crescendo di un paio di minutiā. Il risultato ĆØ stato questo: Orchestral intro. A two-minute emotional crescendo that starts as a breath ā faint strings and airy pads shimmering like dawn light ā and grows into a majestic cinematic overture. Layered violins and cellos in legato rise with warmth; brass enters gradually with bright resonance; timpani rolls, sub booms, and reversed cymbal swells add motion. Synth pads (Juno-60, Prophet-5) and angelic choirs blend with the orchestra, forming a luminous hybrid soundscape. Each 20 seconds, intensity builds ā harp glissandos, choral bursts, and orchestral hits create a sense of awe and anticipation. The tempo accelerates naturally, mirroring a heartbeat shifting from calm to euphoric revelation. Emotion: wonder, elevation, positive tension. Mixing: hi-fi cinematic width, cathedral reverb on strings and choir, dynamic automation following the BPM rise. Goal: an awe-inspiring prelude that transforms silence into radiant energy before the main song begins. Il passo successivo ĆØ stato separare la strumentale dalla voce del brano originale usando LALAL.AI. Ho poi caricato la traccia strumentale su Suno insieme al prompt creato in precedenza, ottenendo esattamente la versione orchestrale che cercavo. Lāultimo step ĆØ stato lāesportazione e la sincronizzazione del prodotto di Suno con la voce originale su Ableton Live, dove ho rifinito i dettagli per ottenere il risultato finale. Lo trovate a questo link: https://lnkd.in/dA8xjxiH Ć stato un esperimento che mi ha fatto capire quanto lāAI possa diventare uno strumento creativo quando la guidi nel modo āgiustoā Il risultato? Ascoltatelo e fatemi sapere che ne pensate! C I A O
-
-
š We were casually browsing Substack the other day and noticed that we were featured in some article as "consumer stem separation app". We found it cute, to be honest, because we branded ourselves as such when we just launched LALAL.AI.Ā At first, we even thought that making karaoke backtracks was one of our primary use cases... Now, five years later though, our solutions are used by audio engineers, live recording engineers, music producers, videographers, SaaS, labels, and large teams in quite complicated scenarios we couldn't even think of when we started. Now, it's:Ā āļøAlmost 7 million usersĀ āļø13M+ hours of audio splitĀ āļøFeatured in Billboard, Forbes, The Times, and many more outletsĀ as well as influencers and celebrities More to go!
-
š¬"Can LALAL.AI remove music from our influencer video and leave only their voice-over?" This is the question our business development team once got. Quick answer: you can! More than that, with LALAL.AI Voice Cleaner, you can: š” Remove any music, like copyright tracks, accidental background music, or post-added sound, from any recording: live, recorded, or post-edited. š” Make your content safe to publish anywhere without copyright strikes. š” Scale it viaĀ API: integrate into SaaS platforms, video editors, or other tools for automatic processing. See how a video with music + voice transforms intoĀ clean, ready-to-publish voice (taken from our own influencer video) š
LALAL.AI ā Professional Voice Cleanup
-
That's exactly the point: you can use several LALAL.AI products in one scenario, as in a bundle. Let's imagine a production team working on a documentary about an iconic music band from the 90s whose original live recordings are far from perfect: the vocals are muddy, the instruments overpower each other, and the sound echoes harshly in the big arena. Here's how LALAL.AI would work as a bundle for this project: š” Sound engineer usesĀ Stem SplitterĀ to isolate vocals and instruments. š” Mixing engineer appliesĀ Echo & Reverb RemoverĀ to clean stadium echo. š” Audio editor runsĀ Voice CleanerĀ on interviews and crowd noise. š”Localization specialist usesĀ Voice ClonerĀ to recreate band membersā voices in other languages. š”Producer / remixer appliesĀ Lead & Back Vocal RemoverĀ to polish tracks for the credits. Outcome:Ā Clean, clear, and authentic audio ready for the documentary release. šTo try all of these, please visit https://www.lalal.ai/
Why is generating audio so complex compared to written words? Words are neat, discrete tokens. We have a long word dictionary and use their indices as numbers to create our embeddings. Audio (like a music sample) is 44,100 floating-point samples per secondāoverlapping voices, reverb, hiss, and instruments all mashed into one āwaveformā. A waveform is simply the raw electrical signal plotted over timeāthe continuous squiggle your mic captures before any processing. Pulling that apart is closer to archaeology than spell-check. To see how AI tackles this mess, letās take LALAL.AI as an example of the āmany small nets > one big netā approach: š§ The mini-model stack: - A Voice Cleaner ā to denoise hums, echo, and leaf-blower ambience so an iPhone mic sounds like a studio. Once this is done, we then have: - A Stem Splitter ā based on their model called Perseus, a transformer trained on multitrack sessions that processes up to 10 stems (vocals, drums, bass, guitars, piano, synths, winds, strings) out of any song. Here, you extract your voice for example. - And, finally, our voice cloner ā learning a speakerās timbre, then re-voices new lines or languages. All thanks to the much simplified and clean track. Each tool is narrow, specialized, and fed its own mountain of dataāthen theyāre stitched together behind one web app/API. Result: you can chain āclean ā clone ā splitā in seconds instead of juggling DAWs and plug-ins. Why AI builders should care (and even follow lalal.aiās approach): Task-specific > monoliths. Smaller nets trained for one job give cleaner outputs and fewer artifacts. Splitting a challenging task into simpler ones, much like splitting a sound into tracks, makes the AI challenge much easier to train and use. The data needed is simpler, tailored to each task, the models can be smaller, the system is more efficient, and you can focus your efforts on the part of the workflow that is truly complicated, like imitating someoneās voice AFTER extracting and cleaning it. Try it directly with your worst Zoom audio or a favourite track and hear the before/after (link in the first comment).
-
-
š¬"I haven't found anything that quite compares to LALAL.AI. I did try an iZotope algorithm once, but it didn't quite live up to my expectations." For Naples-based producerĀ Raffaele Arcella, better known asĀ Whodamanny, precision and emotion are two sides of the same coin. As one of the creative forces behindĀ Periodica Records, heās built a sound rooted in Italyās analog heritage yet open to modern experimentation. In our new conversation with Raffaele, he shares how tools likeĀ LALAL.AIĀ fit naturally into his analog-heavy workflow, helping isolate reference drums, fix mixes on the fly, and keep creativity flowing without ever compromising the soul of the sound.