consider I have a
string1 = "hello hi goodmorning evening [...]"
and I have some minor keywords
compare1 = "hello evening"
compare2 = "hello hi"
I need a function that returns the affinity between the text and keywords. Example:
function(string1,compare1); // returns: 4
function(string1,compare2); // returns: 5 (more relevant)
Please note 5 and 4 are just for example.
You could say - write a function that counts occurrences - but for this example this would not work because both got 2 occurrences, but compare1 is less relevant because "hello evening" isn't exactly found in string1 (the 2 words hello and evening are more distant than hello hi)
are there any known-algorithm to do this?
ADD1:
algos like Edit Distance in this case would NOT work. Because string1 is a complete text (like 300-400 words) and the comparing strings are max 4-5 word.
py-editdist
will give you the Levenshtein edit distance between two strings, which is one metric that might be helpful.See: http://www.mindrot.org/projects/py-editdist/
The code example from that page:
Related: https://stackoverflow.com/questions/682367/good-python-modules-for-fuzzy-string-comparison
Take a look into creating N-grams out of your input data and then matching on the N-grams. I have a solution where I regard each n-gram as a dimension in a vector space (which becomes a space of 4000 dimensions in my case) and then affinity is the cosine of the angle between two vectors (the dot-product is involved here).
The hard part is to come up with a metric defining the affinity in a way you want.
An alternative is to look at a sliding window and score based on how many words in your compare_x data is in the window. The final score is the sum.
Well, you can count the occurrences of pieces of the comparing text, ie:
"a-b-c" -> "a" , "b" , "c" , "a-b" ," b-c" , "a-b-c" (possible "a-c", if you wanted that)
And then count occurrences of each of those, and sum them, possibly with a weight of (length of string) / (length of whole string).
Then you just need a way to produce those pieces, and run a check for all of them.
While the Levenshtein distance as it stands may not suit your purposes, a modification of it might: Try implementing it by storing the insertions, deletions, and substitutions separately.
The distance will then be the sum of the following:
You'd have to test this, of course, but if it doesn't work well try simply using the sum of consecutive insertions/deletions (so, " hi good morning " is only 1 edit).
EDIT
P.S.: this assumes a relatively major change to how Levenshtein works, you'd want to 'align' your data first (finding out where there's significant (more than two characters) overlap and inserting 'null' characters that would count as insertions).
Also, this is just an untested idea, so any ideas for improvements are welcome.
Here you can find a list of metrics to calculate distance between strings, and an opensource java library that just do that. http://en.wikipedia.org/wiki/String_metric In particular, take a look at the Smith–Waterman algorithm, keeping in mind that what they call "Alphabet" can be composed by what we call Strings : so, given the alphabet
and called d the distance, your function tries to calculate
A Dynamic Programing Algorithm
It seems what you are looking for is very similar to what the Smith–Waterman algorithm does.
From Wikipedia:
Let's see a practical example, so you can evaluate its usefulness.
Suppose we have a text:
I isolated the segment we are going to match, just for your easy of reading.
We will compare the affinity (or similarity) with a list of strings:
I have the algorithm already implemented, so I'll calculate the similarity and normalize the results:
Then we Plot the results:
I think it's very similar to your expected result.
HTH!
Some implementations (w/source code)