4pr0n / rip Goto Github PK
View Code? Open in Web Editor NEWalbum ripper
License: GNU General Public License v2.0
album ripper
License: GNU General Public License v2.0
Your site is great, but are you considering allowing downloads from http://www.gallery-dump.com?
Thank you for your work.
Some tumblr posts contain a series of pictures.
Every time I try to [rip the album] it ends up downloading everything on the tumblr instead of just the one post.
What setup is required to make gone wild ripper work? I put a request in with gonewild:username in my locally hosted version of this but it always fails.
I'd really love it if rip and zip on deviantart.com included the scraps tab, or at least the scraps could be entered as a URL. Currently entering the scraps URL only collects the main gallery.
Current system deletes based on FIFO:
Viewing or ripping an album will reset the last-updated time to the current time.
This means least-recently-accessed albums will be deleted first.
Deletions occur when total size of all /rips/ gets above a threshold (currently 150GB).
Ideally:
I see you've managed to get reddit users, so how about subreddits?
However I do believe the subreddit.json only allows for the 1000 most recent posts
Pretty simple. Mainly occurs with imgur 'selections' like imgur.com/a/E49f1#6AG3G (I made that url up just now btw, so it probably doesn't point at anything.).
The exact effect seems to be that instead the download button pointing at rip.rarchives.com/#ESCAPED_URL (in my fictional example this would be rip.rarchives.com/#imgur.com%2Fa%2FE49f1 -- well, this is the form that seems to work, though the full escaped url would also include the (escaped) #6AG3G), it points at rip.rarchives.com/#ESCAPED_PART_OF_URL_AFTER_HASH (in my fictional example this would be rip.rarchives.com/#6AG3G).
Naturally this totally fails to recognize the url being sent in as anything valid.
When trying to download an album off deviantart a lot of images in the zip end up as 1 kb broken files.
A small percentage of the users are consuming a large percentage of the bandwidth and resources.
The system requires grepping access logs to find offenders & manually banning their IPs.
This could be a cron script (or embedded in python_cleanup.py). Look for users ripping albums over a threshold and ban them.
For ex, hand out bans for users that rip more than 50 albums in one day.
Bans could be reset at the end of the day, or after 2 days.
Banning should redirect to a page informing the user that they've been banned and [may] be unblocked in the near future.
I'd really love it if imgsrc.ru albums could be ripped by rarchive's ripper. That would be very helpful for me. Awesome ripper, fantastic work.
Any change we can add support for your own site? :) This would be perfect for user rips.
Support this image site: http://i.rarchives.com/?url=user:Kaitlin328
Got a different request earlier today.
It would be trivial to use the api.
And i.rarchives could link back to rip.rarchives.
Whether it is in one shot or the ability to include what page number/submission range to download, it would be very nice. It the largest galleries that I'm going to want to use this the most for.
Something like http://tyilo.imgur.com/all/ or just http://tyilo.imgur.com/
I see that it's capped at 500 images...
If you were to introduce some kind of pay system that could eliminate that limit... I'd give you my money, baby.
I can't think of an easy way to monetize the site, so I'll work on making the ripper executable from the command-line easily. That way people can rip to their heart's desire on their own desktop.
Gotchas:
basesite.py is uuuugly. Showing how to use the built-in features could be handy.
This will require updating _testsite.py to use the latest features (urls_only, create_thumb, etc).
Similar to the warning used on i.rarchives.com; check cookie & do not display content until 'i agree' is clicked.
Also, while you're at it, add a disclaimer about the shelf-life of the zips/albums, and how the caching works.
Here's a fun one for you: when middle-clicking on the recently downloaded links,
- the first opens in a new tab and keeps focus on the primary site.
- The second, however, opens a new tab and sets focus to it.
- All tabs after that follow the same procedure.
This changed after the latest revision. No big deal, just a fun line of code gone awry somewhere!
Any chance you could add support for facbook and rajce www.rajce.idnes.cz? Flickr rip doesn't seem to work when trying to rip an album. The firefox extension could use an update for the new supported sites. Thanks for all your hard work in making this epic ripper.
On systems, such as Arch Linux, where Python 2 and 3 can coexist, it's a mistake to refer directly to the 'python' binary in a shebang: this will not always point at Python 2. On Arch for example, it points at Python 3, which 'rip' is not compatible with.
However, 'python2' can be relied on to always point at Python 2.x :) See http://www.python.org/dev/peps/pep-0394/
With that change to the shebang line, I was able to successfully install 'rip' on (Lighttpd + mod_cgi / Arch Linux x86_64)
Most sites don't let you download the .flv or .mp4 file directly. Having a box to paste the video site & get back a downloadable link to the video would be a nice feature.
Example: schoneseele.tumblr.com/tagged/best_of_ME
Likely due to the way site_tumblr.py parses the working directory in download()
from primoarnold sent 8 hours ago
I have a request. Is it possible to get the image ripper/zipper to work with http://society.pichunter.com/ The site requires a login but is free and is free of malware/spyware and adds as far as I can tell. Thank you in advance for your consideration in this.
from 9959 sent 5 days ago
I have a suggestion: Support for Sankaku chan! :D http://chan.sankakucomplex.com/
from Saiboogu sent 16 days ago
Suggestion - Capture captions somehow. Maybe just toss them in the log.txt or make a caption.txt with image file names and caption text.
Download media from twitter search results i.e #cosplay
Is there any way you could add a feature to rip all image galleries off of a user's reddit page with your album ripper?
Right now you have to open each gallery or right click->copy link location and then paste into the ripper.
At the moment it appears that the flickr downloader downloads only a "large" version of the files but it is often possible to download a higher res (original) copy of the image.
For example the files in this folder, http://www.flickr.com/photos/lukejohnson/sets/72157622361782148/ which become downloaded to 640 x 1024
instead of 1200 x 1920
I'm not sure what is going on here, but it seems that it can't find the directory for putting the rip in, even if I create the directory it wants, it deletes it (on linux), and refuses to find it. I've tried on both windows and linux.
A declarative, efficient, and flexible JavaScript library for building user interfaces.
๐ Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. ๐๐๐
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google โค๏ธ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.