This project employs Python3 and BeautifulSoup to scrape a Tumblr site (with the url provided by the user) to automatically download, page by page, all the image files (JPEG, GIF, PNG) from the site's posts. This is an excellent tool for archiving other peoples' Tumblrs <3
HOW TO USE (for non-programmers)
Make sure you have room on your drive for what could be several gigabytes of images and GIFs!!!
Okay,
OPEN TERMINAL -or, for non-programmer folks:
- On your Mac keyboard, hit Command + [SPACEBAR]
- type "terminal" and hit the enter key (you should now see a black field that allows you to type things) https://www.wikihow.com/Open-a-Terminal-Window-in-Mac
https://www.quora.com/How-do-I-open-terminal-in-windows
python get-pip.py
sudo apt-get python 3.6.1
sudo pip3 install beautifulsoup4
- And the script: download Tumblr_Image_scrape.py file from this github repository https://github.com/nigeld3v/Tumblr_Image_scrape/blob/master/Tumblr_Image_scrape.py
In Terminal, you want to change your working directory to the folder where you have downloaded 'Tumblr_Image_scrape.py' If you do not know how to navigate the files on your computer using Terminal, here is a good resource:
You should easily be able to find the right file using "pwd", "cd", and "ls" commands in terminal.
Once you have changed your directory to the file where Tumblr_Image_scrape.py was saved, type:
Python3 Tumblr_Image_scrape.py
and then press the [Enter] key on your keyboard.
You will be prompted to provide a Tumblr URL. Type or paste the address of the Tumblr, then press the [Enter] key on your keyboard.
Press Enter. The script should run and download the Tumblr site's images, page by page, and store them in a folder (title will be the Tumblr site's name) in the same directory where Tumblr_Image_scrape.py is kept.
NOTE: if it works, this script will continue to run until you make it stop. The easiest way to do this in Terminal is to hit the [Control]+[C] keys on your keyboard.
<3 <3 <3 <3 <3