- Ability to scrape users on flickr based on user tags or ids of flickr groups Note: user_ids have been filtered to those who have been on Flickr for > 2 years and who have at least 1000 photographs.
- Saves user ids into a .txt file, user information into a .json and .csv file
- Ability to scrape all user's photos from a .txt file of user ids.
- Ability to restrict photos scraped to those that are geotagged only
- Ability to download all photos scraped into specified folder
- Ensure that the relevant modules have been installed via
pip install <insert_module>
Some key modules include: flickrapi, datamuse, geopy Note: To import datamuse, usepip install python-datamuse
- Start by adding your flickr API and secret key into the respective python files.
- Run the respective files via
python <insert_file_name>
- Run flickr_scrape_users to make use of Features 1 and 2
- Run flickr_retrieve_all_photos to make use of Features 3 and 4
- Run download_images to make use of Feature 5
- Runtime for large amounts of data (~500 to 1,000 user ids) will take a significant amount of time and even longer if the free geo-coding service is used
- To use the paid geo-coding service (I.e. Google Maps), please ensure that you have a Google API key. Else, Nominatim will be used.