Meta Segment Anything Model Audio

ai.meta.com

200 points by megaman821 3 days ago


blagie - 2 hours ago

For future ML developers: A post like this should include system requirements.

It's not clear from the blog post, the git page, and most other places if this will run on, even in big-O:

* CPU

* 16GB GPU

* 240GB server (of the type most business can afford)

* Meta/Google/Open AI/Anthropic-style data center

hbn - 12 hours ago

I hope we keep making progress in isolating tracks in music. I love listening to stems of my favorite songs, I find all sorts of neat parts I missed out on. Listening to isolated harmonies is cool too.

locusofself - 10 hours ago

As someone recording myself playing music, I've been meaning to see if any of these tools are good enough yet to not only separate vocals from another instrument (acoustic guitar for example), but do so without any loss of fidelity (or least not a perceivable one).

The reason I'm interested in this is because recording with multiple microphones (one on guitar, one on the vocal), has it's own set of problems with phase relationship and bleed between the microphones, which causes issues when mixing.

Being able to capture a singing guitarist with a single microphone placed in just the right spot, but still being able to process the tracks individually (with EQ, compression, reverb, etc), could be really helpful.

ortusdux - 13 hours ago

Would be great for the hearing impaired and CAPD sufferers when combined with Meta glasses or the like.

kace91 - 8 hours ago

Funny that:

- This feature is awesome for sample-based music

- Sample music is not what it was due to difficulties related to legal rights

- This model was probably created by not giving a damn about said rights

moss_dog - 5 hours ago

This is incredible! I wouldn't have thought it was possible to cleanly separate tracks like that. I wonder to what extent the model is filling in gaps, akin to Samsung's "ultra zoom" moon.

tasty_freeze - 9 hours ago

I use moises frequently for track separation for learning songs. It does pretty dang well. I was shocked that the score of moises is ranked way worse than just about everything else, including lalal.ai, which I also used before buying moises. Perhaps lalal.ai has gotten better since I last tried it.

cyberax - 8 hours ago

Can this be used to nuke the laugh tracks?!?

htrp - 2 days ago

super amazing demo performance being able separate out music voice and background noises. do you have to explicitly specify what type of noise to separate?

mwmisner - 12 hours ago

Playing with the background I tried to Isolate just the espresso machine and the train sounds in one of their demos and it seemed to fail. Maybe not the desired use case, but I thought it was odd that I could break it so easily on the sample material.

Escapade5160 - 8 hours ago

From my brief testing in the playground, it is not very good. Maybe it needs better prompting than the 1 word examples.

nmstoker - 7 hours ago

Would be interesting to leverage the non spoken/environment noises to guide what level of detail and style of speech a chatbot replied with, for instance being more casual, gentle, with a touch more detail if in a quiet home/office environment, but more curt and concise with emphasized diction if the person is traveling, such as in a noisy train concourse. People tend to do that subconsciously but bots ignorantly wittering on can be annoying and hard to use because they miss the cues.

brador - 3 hours ago

How about they fix their MusicGEN model on hugging face first.

almosthere - 11 hours ago

mSAMA haha, get it

emsign - 2 days ago

[flagged]