I also recommend use of agrep. The only caveat is the restrictions on free use for commercial applications. I don't believe there is anything more efficient or better suited. The link that tachyon and I point to has links to other libraries and applications.
| [reply] [Watch: Dir/Any] |
From a fairly quick perusal of the options, I don't think agrep will help much, except maybe as a pre-filter.
- It won't report where in a line a a match was found.
- It stops matching against a given line when it finds the first match.
- If you supply a file of things to match, it doesn't tell you which one matched.
Maybe I missed some things in amongst the six help 'screens'?
Examine what is said, not who speaks.
"Efficiency is intelligent laziness." -David Dunham
"Think for yourself!" - Abigail
"Memory, processor, disk in that order on the hardware side. Algorithm, algorithm, algorithm on the code side." - tachyon
| [reply] [Watch: Dir/Any] |
I went to the University of Arizona, and as an undergraduate, I would sit in on masters- and postdoc- level classes where folks were discussing stuff like agrep.
Somewhere in my files I have the follow-up to this paper, which allows for affine weighting for various symbols. For example, you might say that vowels are more interchangeable than consonants, if you're looking for fuzzy matches in the pronunciation problem space. I think the author of that paper went on into bioinformatics in a big way after that.
-- [ e d @ h a l l e y . c c ]
| [reply] [Watch: Dir/Any] |