![]() |
#1 |
THE NOOB
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() Posts: 701
Karma: 1545649
Join Date: Jan 2012
Location: Italy
Device: Kindle Touch 5.3.2
|
Wget & recursive download
Good afternoon!!!
I'm here to ask a noobish question (once again)... I'm using Code:
wget Code:
-r Is there a workaround? Thanks to everyone!! ![]() |
![]() |
![]() |
![]() |
#2 |
Carpe diem, c'est la vie.
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() Posts: 6,433
Karma: 10773668
Join Date: Nov 2011
Location: Multiverse 6627A
Device: K1 to PW3
|
You could use a standalone "full featured" wget. There may be one in twobob's collection...
|
![]() |
![]() |
Advert | |
|
![]() |
#3 |
THE NOOB
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() Posts: 701
Karma: 1545649
Join Date: Jan 2012
Location: Italy
Device: Kindle Touch 5.3.2
|
I tried downloading this wget, I extracted the package and placed in /mnt/us.
Then I tried: Code:
/mnt/us/wget http://www.google.com Code:
-sh: /mnt/us/wget: not found |
![]() |
![]() |
![]() |
#4 |
Going Viral
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() Posts: 17,212
Karma: 18210809
Join Date: Feb 2012
Location: Central Texas
Device: No K1, PW2, KV, KOA
|
Hmm...
/mnt/us is FAT32 - - Doesn't FAT32 support execute permission? Do an: ls -l /mnt/us/wget And see if it is marked executable. Also do a: mount and see if the mount parameters given when mounting the FAT32 was given the "exec" permission or was given the "noexec" permission. Note: My fingers and my keyboard are not getting along today. Always wait at least 10 minutes after this post time before quoting or replying - it can take me that long to spot errors and typos. |
![]() |
![]() |
![]() |
#5 | |
Carpe diem, c'est la vie.
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() Posts: 6,433
Karma: 10773668
Join Date: Nov 2011
Location: Multiverse 6627A
Device: K1 to PW3
|
EVERYTHING on the USB drive "automagically" has execute permission (not a FAT attribute, so assumed).
Quote:
http://packages.debian.org/sid/armel/wget/download It is the same URL that you provided (http://packages.debian.org/sid/armhf/wget/download), but with "armhf" changed to "armel". Try that one. ![]() Last edited by geekmaster; 09-21-2012 at 11:46 AM. |
|
![]() |
![]() |
Advert | |
|
![]() |
#8 |
THE NOOB
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() Posts: 701
Karma: 1545649
Join Date: Jan 2012
Location: Italy
Device: Kindle Touch 5.3.2
|
Thanks Baf, that wget works really well!!!
It's a good starting point for my future WAF (if it will ever become real...) |
![]() |
![]() |
![]() |
#9 |
( ͡° ͜ʖ ͡°){ʇlnɐɟ ƃǝs}Týr
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() Posts: 6,586
Karma: 6299991
Join Date: Jun 2012
Location: uti gratia usura (Yao ying da ying; Mo ying da yieng)
Device: PW-WIFI|K5-3G+WIFI| K4|K3-3G|DXG|K2| Rooted Nook Touch
|
yeah I was gonna point him at bafs. my work is already done.
There is another one linked to a static curl ligging about somewhere IIRC also. Baf's will do the job fine (@ GM I didn't even bother building one, nor a locate, nor an updatedb, nor anything baf had already done. his versions are shiny and tidy, I use them) |
![]() |
![]() |
![]() |
#10 |
THE NOOB
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() Posts: 701
Karma: 1545649
Join Date: Jan 2012
Location: Italy
Device: Kindle Touch 5.3.2
|
Ok, here I am again...
Baf's wget works really fine (thanks a lot!!!) but I really don't know how to set it up nicely... Here the situation: I would like to download a web page (single web page) from Wikipedia, so I run this: Code:
export PAGES="/mnt/us/WebPages" webpage="http://it.wikipedia.org/wiki/Wget" /mnt/us/extensions/offlinepages/bin/resource/Wget/wget -e robots=off --wait=20 --limit-rate=20K --quiet --no-parent --page-requisites --convert-links --adjust-extension -U Mozilla -P "$PAGES" "$webpage" BUT it doesn't download images... It retrieves only Wget.html and favicon.ico!! Isn't it supposed to download all the necessary to display the page? If I turn on the --recursive, I get all the linked html to the wget page but no images again!! Any help? Thanks guys!!! ![]() |
![]() |
![]() |
![]() |
#11 |
Going Viral
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() Posts: 17,212
Karma: 18210809
Join Date: Feb 2012
Location: Central Texas
Device: No K1, PW2, KV, KOA
|
Images are probably protected from off-site references.
Try having wget pass that web-site's set of cookies so that it thinks the image references are coming from a browser viewing that site. You might also have to fake up a "refer" header for wget. Both described in the wget documentation available - I don't know the details off hand (I would have to look them up like any other mere moral.) |
![]() |
![]() |
![]() |
#12 |
THE NOOB
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() Posts: 701
Karma: 1545649
Join Date: Jan 2012
Location: Italy
Device: Kindle Touch 5.3.2
|
Thanks for the hint...
I tried this: Code:
/mnt/us/extensions/offlinepages/bin/resource/Wget/wget --random-wait -e robots=off --limit-rate=20K --page-requisites --no-parent --convert-links --adjust-extension --restrict-file-names=windows --cookies=on --load-cookies=/var/local/waf/browser/cookieJar --keep-session-cookies -U mozilla -P "$PAGES" "$webpage" I even tried with: Code:
--referer=http://it.wikipedia.org ![]() |
![]() |
![]() |
![]() |
#13 |
( ͡° ͜ʖ ͡°){ʇlnɐɟ ƃǝs}Týr
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() Posts: 6,586
Karma: 6299991
Join Date: Jun 2012
Location: uti gratia usura (Yao ying da ying; Mo ying da yieng)
Device: PW-WIFI|K5-3G+WIFI| K4|K3-3G|DXG|K2| Rooted Nook Touch
|
hmm. possibly do a bit of playing around with fiddler then. in firefox.
|
![]() |
![]() |
![]() |
#14 |
THE NOOB
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() Posts: 701
Karma: 1545649
Join Date: Jan 2012
Location: Italy
Device: Kindle Touch 5.3.2
|
Ok, here I am...
The problem when "wgetting" a page from wikipedia is that you need to turn on recursive download as the normal "--page-requisites" option doesn't retrieve images (it gets only the html to be honest). This is because images are hosted on a separate page! BUT turning on "--recursive" gives you LOTS of things to download (as wiki pages are full of external & internal links). So I ended up using --recursive + --span-hosts + --domains. This is the complete command: Code:
wget --recursive --span-hosts --domains=upload.wikimedia.org -e robots=off --random-wait --limit-rate=20K --page-requisites --no-parent --convert-links --adjust-extension --restrict-file-names=windows -U Mozilla -P "$folder" "$webpage" I'm going to investigate where the hell is it hosted... |
![]() |
![]() |
![]() |
#15 | |
( ͡° ͜ʖ ͡°){ʇlnɐɟ ƃǝs}Týr
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() Posts: 6,586
Karma: 6299991
Join Date: Jun 2012
Location: uti gratia usura (Yao ying da ying; Mo ying da yieng)
Device: PW-WIFI|K5-3G+WIFI| K4|K3-3G|DXG|K2| Rooted Nook Touch
|
Quote:
|
|
![]() |
![]() |
![]() |
|
![]() |
||||
Thread | Thread Starter | Forum | Replies | Last Post |
Could anyone compile full-feature wget for Touch | wisecrick | Kindle Developer's Corner | 15 | 09-09-2012 10:41 AM |
wget an ftp folder | mephis | PocketBook | 3 | 10-11-2011 05:34 PM |
902 wget pdf newspaper from site? | niroma | PocketBook | 23 | 07-10-2011 03:09 PM |
How to use wget to download an online HTML book | amoroso | Lounge | 11 | 04-25-2011 05:10 AM |
Recursive conversion of Websites to Kindle | oecherprinte | Amazon Kindle | 0 | 10-18-2010 07:22 AM |