So, not really a linux question persay, however, I'll probably be trying to implement this with perl under linux. :) Say I have a very large file of random characters, like 8GB. And I have another smaller file of random data. I want to take the largest chunks I can from the small file, and find out where in the large file they will fit (match). Statistically, how likely is it to match strings of length 50 chars, 100, 200, etc... Also, what kind of search algorithm would be best for this. Say I'm trying to match a string 50 characters long to something in the larger file and it matches, but, if I had started that string 20 characters sooner, I would have been able to match 70 characters instead of just 50. I want kind of a fuzzy search algorithm that can find the largest matching pieces first, and then match the smaller leftovers. Any ideas? :) Jay _______________________________________________ TCLUG Mailing List - Minneapolis/St. Paul, Minnesota http://www.mn-linux.org tclug-list at mn-linux.org https://mailman.real-time.com/mailman/listinfo/tclug-list