Montag, 8. Juli 2013

Download all PDF files from Springer Link or other websites in Ubuntu using wget



As a regulary visitor of Springer Link I was always bored to download every single chapter of a publication. So I put together a small wget-comment which allows me to automatically download multiple PDF files from a webpage:


wget -nd -e robots=off --wait 0.25 -r -A.pdf http://yourWebsite.net/


The option -nd stands for no-directories, which means no directories are downloade and -e robots=off forces to not download the robots.txt, because we only want PDFs. The wait command ensures that failed downloads are retried after 0.25 seconds. Additionally I use the r parameter to get PDFs from subpages recursively and -A.pdf to get PDF files only.

Got improvements ? Leave a comment !

Keine Kommentare:

Kommentar veröffentlichen