I wished to download ALL links beginning with
and NO OTHER LINKS which do not start with the name of the above url.
In fact when I tried to download, by simply giving the above url in the new project option along with default settings, not only did i get webpages not starting with this url but I also did not get all the webpages starting with this url. I had to manually check the 100 pages to find out the fifteen web pages which were not downloaded.
Also, I am not good at understanding filters and parsing. So I will be very grateful if you could specify the exact configuration that I need to make before starting to run the project.
Though the Project looked perfect, unfortunately it failed to work.
I could only download the following :
There are all files from /gdebgb1 to /gdebgb100 atleast. Infact, the above list shows the urls that were in the queue. Urls other than these never even entered the queue. Also, whwnever I click on update or restart the projecti get this in the download progress window :
Download Complete.Status 302 object moved. I even went to the website to see if all the pages were still there and i found that they were there.
I would be glad if you could try to download all the pages once,bcoz in all they are 100 pages of about 20KB each----which means total 2MB.
However I can't understand that why only certain urls were downloaded like the list i sent u before. After all, the rest (which were not downloaded) also did not seem to be linked inside a link.
(i mean why require a different setting for /gdebgb94 and /gdebgb2 ?) How did u determine that level limit should be increased? I ask u this bcoz the answer may help me in future projects.
Why is there no error reporting about partially downloaded links?
Is there a way to find these partially downloaded links and have them automtically download themselves?
Thanks for all the prompt replies. It was almost like I was in an online chat with you. I have never seen such a fast response anywhere. Thanks a lot.