This Python scraper, built with Scrapy, takes country URLs and dives deep into beach-searcher pages to uncover all its beaches. It follows the provided country link, meticulously gathers beach details, and stores them neatly in a JSON file, ready for further analysis.
- Extracting Complex Listing Structure and Extract Dynamic Prices
- Parse API responses and then request further child pages
- Implemented Download Delay to avoid server overloading.
Python 3 (Linux/Mac/WSL/Windows)
To install dependencies open terminal and type:
git clone https://github.com/seemab-yamin/beach_searcher_spider/
cd beach_searcher_spider
pip install -r requirements.txt
First update config.json to add target urls in "start_urls" list in JSON dict.
python3 main.py
- Scrapy: https://scrapy.org/