"J. Linder" <jann@jann.com> writes: }funny...I was kinda looking forward to an automated system. There are over }500 pages on our site and over 1000 images in 42 directories. } }i would just like a program to read in all the .htm, .pl, .cgi, .html, }.shtml extensions and i can parse them for the .gif and .jpg's } }then i could make a list that we could work from to manually delete the }unused files. Sounds like you want a site checking robot. I know of two possibilities. The first is a shareware program, Big Brother, <http://pauillac.inria.fr/~fpottier/mac-soft.html.en>. I use it as a simple link checker for my bookmark files, but it has the ability to recurse, so it should be possible to use it as a site checker. The second is Checkbot, <http://dutifp.twi.tudelft.nl:8000/checkbot/>, which is written in Perl and requries libwww-perl-5. I have no idea whether it works under MacPerl or not, but with enough effort you should be able to get it working. At the very least it should show you the way to write your own robot. } }The problem is i dont know how to make a listing of all the files on my }site and cycle through each filename in order to parse them using perl. }The parsing is easy! } }thanks } }jann } } --- Paul J. Schinder NASA Goddard Space Flight Center Code 693, Greenbelt, MD 20771 schinder@pjstoaster.pg.md.us ***** Want to unsubscribe from this list? ***** Send mail with body "unsubscribe" to mac-perl-request@iis.ee.ethz.ch