tumblr-downloader
A tumblr scraper, designed to download all the images from the blogs that you want.
##Features
- Photo, video, and audio downloading
- Iterative downloading -- If you download from a blog, the run it again, only the missing downloadables can be downloaded the second time.
- Complete downloading -- Will scan the entire blog for downloadables, not just the first X pages.
- Rate limiting
- Concurrency -- download from multiple blogs at the same time
Download
Latest releases can be found here for Windows, Mac, and Linux.
If you are willing to help improve this program, please download the debug
version. It takes a bit more RAM while downloading, but its output will help track down bugs. Other than that, it has no change, and will still download everything that the normal version does, and at the same speed.
Usage
###Simple
Make a text file called download.txt
with each tumblr blog you want to download on a separate line:
nature-pics
sunsets
chickenpictures
Run tumblr-downloader
once it's complete. It'll download all the pictures from the blog and save it in a downloads/<username>
folder for each user.
You can also download a single tag for a blog, if you only want specific content. For example, you can have the following:
nature-pics forests
sunsets
chickenpictures funny faces
If your tag has spaces in it, just type the tag normally after the blog name. For instance, in the above example, chickenpictures
will download anything tagged with funny faces
. (Note that it will NOT download funny
and faces
separately like this.)
Command line
Run tumblr-downloader
as such, appending the usernames you want to download after the executable:
$ ./tumblr-downloader nature-pics sunsets chickenpictures
Command line options
-d
- The maximum number of images to download at the same time. Default is 3.
-r
- The maximum number of requests per second to make.
-f
- Force check -- the downloader will recheck old tumblr posts to see if it missed anything.
-dir
- The directory to save files in. By default, tumblr-downloader saves the files in the same directory it's run from, making a new directory for each blog.
-ignore-audio
, -ignore-videos
, -ignore-photos
- Skips downloading the respective types of files.
-p
- Enable progress bar to track progress instead of printing files being downloaded.
-server
- Runs as a server, automatically restarting the download process after finishing.
-sleep
- Only works if -server
is enabled. The amount of time to wait between download sessions.
Suggestions
Use the issues
tab provided by Github at the top of this project's page.
Contributing
- Fork it!
- Create your feature branch:
git checkout -b my-new-feature
- Commit your changes:
git commit -am 'Add some feature'
- Push to the branch:
git push origin my-new-feature
- Submit a pull request :D