[Date Prev][Date Next][Thread Prev][Thread Next] [Search] [Date Index] [Thread Index]

Re: [MacPerl-WebCGI] Indexing a remote site



robinmcf@altern.org skrev/wrote:
>I don't know if anyone else is annoyed by this, but recently I've 
>been rooting about in the PDF versions of "Inside Macintosh" 
>(http://developer.apple.com/techpubs/mac/pdf/), on previous visits I 
>downloaded some zipped bundles of docs. Apple seems to have changed 
>it's way of doing things since then and now you view them on-line, 
>not a convenient option if you have to pay for telephone time. So I 
>wrote the script down below using LWP (which to my total surprise, 
>and to the authors Gisle Aas & Martijn Koster 's credit, worked 
>first time)  to download the PDFs onto  my hard disk. So far no 
>complaints, _but_ , I tried (and failed) to grep a  file list which 
>I hoped to pass to the rest of the script  but ended up having to 
>compile the list of docs I wanted to get by hand. I couldn't seem to 
>use READDIR() and LPW only fetches documents. It goes without saying 
>that I don't have ftp or telnet access to the directory, but is 
>there a way to generate a file list on the fly?

You could get the list
$pagewithlinks = get("http://developer.apple.com/techpubs/mac/pdf/");

and parse that for filenames...
Christian

==== Want to unsubscribe from this list?
==== Send mail with body "unsubscribe" to macperl-webcgi-request@macperl.org