Skip to main content

Regarding grabbing images from Google (such those linked in the G+Exporter files)

Regarding grabbing images from Google (such those linked in the G+Exporter files)

Are there known tools around that can mass download images from URLs without triggering any kind of throttling or denial from Google's servers?

Comments

  1. I use JDownloader. Works surprisingly good for this purpose.

    ReplyDelete
  2. Martina Neumayer My Chrome is not happy about the .rar file for JDownloader and says "Virus detected".

    ReplyDelete
  3. Lars Fosdal Well known false positive on windoze machines. I use Arch so I don't care about some viri-alerts ;)
    Here is an official installer source:
    jdownloader.org - JDownloader.org - Official Homepage

    ReplyDelete
  4. Martina Neumayer That's the page I used. It redirects to a Mega.nz download - which also warns about a virus.

    ReplyDelete
  5. Lars Fosdal Yeap..the charms of having a windows system onboard ;)
    Btw.. As the support stated on their website:
    "..The installers on the main page asks you if you want to install a bundled software - for example a toolbar. This is the way we earn money. The installer shows you a screen, and you can uncheck the toolbar installation. So you you read the installer screens, you will not get anything unwanted installed.

    We do not like this Adware-Driven model alot, but there is no real alternative right now. We will offer a kind of "Paid-JD-Premium" Version of JDownloader that has no ads, and a few extra features. But this Premium services will come when we finally release JD2 (it is beta right now)

    Unfortunately, JD2 Development is slower than we thought, and so we offer Adware-free installers here until we finally prepare for JD-Premium.".

    That's why this file can trigger a warning. Just uncheck this additional option and you're good to go.

    One more thing.. You can use also this "jar" binary file. Just rpm on it, choose "open with.." and then select to open it with "Java(TM) Platform SE binary". Should works without problems.

    ReplyDelete
  6. The likelyhood of me installing something that has been flagged - correctly or incorrectly = zero. Call me paranoid, but 35 years of computing has made me quite careful.

    ReplyDelete
  7. Lars Fosdal Careful and youre using windoze? hahaha good one.. You made my day xD

    ReplyDelete
  8. Martina Neumayer Ridiculing an OS = so mature.

    ReplyDelete
  9. Lars Fosdal Nope.. not os. And mature is reading a text with understanding. ;)

    ReplyDelete
  10. Lars Fosdal My occasional G+ web scraping of 3k - 50k URLs at 1s intervals has never triggerred throttling AFAIA.

    We might be helping someone meet their KPIs!

    ReplyDelete
  11. Edward Morbius Thank you - that sounds promising.

    ReplyDelete
  12. Lars Fosdal Yeah. I've tried to avoid tempting fate by generally including a 'sleep 1' or 'sleep 2' term in my shell loops. But I've never seen limiting occur.

    I think Google may impose a rate limit on their side via QoS throttling or similar mechanisms to about 1 query/s, based on behaviour whilst I've tested such scripts on small (10-100 query) runs.

    ReplyDelete

Post a Comment

New comments on this blog are moderated. If you do not have a Google identity, you are welcome to post anonymously. Your comments will appear here after they have been reviewed. Comments with vulgarity will be rejected.

”go"