At 21:11 -0400 4/12/1999, Kevin Reid wrote: >I once wrote a search engine that searched a large tab-delimited text >file. The way I handled limiting results was that I simply stopped >searching after N hits, and put a link at the bottom that included the >byte position in the file at which the search stopped so that the search >could resume from that point. > >The limitation of this method is that the results appear in their order >in the file. That would probably be acceptable if I just had a single document to deal with, but I've got around 4000 pages of stuff scattered over 3 directories in maybe 70 files. It's beginning to look like maybe you first write out a file with all of the hits, then open it and send the first 20 along with a link back into the file that is supposed to extract the next 20 or something. This seems pretty messy since I guess you'd have to use some kind of serialization scheme to name the files to keep them straight and you've still got to get rid of all of them at some point. With a stateless browser connection, I don't know of another way to avoid having to conduct a new search from scratch. Thanks. Richard Gordon -------------------- Gordon Consulting & Design Database Design/Scripting Languages mailto://maccgi@bellsouth.net 770.565.8267 ==== Want to unsubscribe from this list? ==== Send mail with body "unsubscribe" to macperl-webcgi-request@macperl.org