I'm developing an iOS application with SDK for iOS 5.0 and XCode 4.2.
I want to develop an application that recognize sounds. I see there is an application called Sound Hound that recognize music and tells artist and title.
How can I do something similar? I want to compare a sound to an existing sound database. How can I do that?
Maybe I can use Fourier Transform. I don't know how to process sounds. Or it could be similar to speech recognition, isn't it?
Shazam Application is the one of the best Example for Application in Linked open data it takes the short sample music from the end user and identifies the song from datasets as well as link to purchase the album.
The user tags a song for 10 seconds and the application creates an audio fingerprint based on some of the anchors of the simplified spectrogram and the target area between them.
For each point of the target area, they create a hash value that is the combination of the frequency at which the anchor point is located, the frequency at which the point in the target zone is located, and the time difference between the point in the target zone and when the anchor point is located in the song.
Once the fingerprint of the audio is created, Shazam starts the search for matches in the database. If there is a match, the information is returned to the user; otherwise it returns a “song not known” dialogue.
I came across a paper which explains how audio search algorithms work. Here is the link. It was written by one of the developers of Shazam, a rival application of SoundHound.
good links on the wikipedia page include: http://www.soyoucode.com/2011/how-does-shazam-recognize-song and the paper http://www.ee.columbia.edu/~dpwe/papers/Wang03-shazam.pdf Sub_stantial noted earlier