- Webscraping
- Data cleaning and transformations
- Writing files locally and to the cloud
- Using an orchestration tool
- Python
- Postgres
- Docker
- Google Cloud Storage
- Prefect (orchestration)
- Fetched all this info from each listing in the selected suburbs + the url of the listing
- End up with hundreds of datapoints like this:
- After cleaning they look like this, the formats are now usable for analysis
- Create real estate price indexes at the province, country or suburb level
- Automate searches for good deals for personal or investment purposes
Collect any sort of structured data from the web at set intervals and use it for personal or commercial purposes
Message me and I'm happy to run through the code with you or help you adjsut it for your purposes Otherwise I tried to make the functions readable. So start at the final_scraper.py file and read through it.
- Dockerize it all so that it can be more easily collaborated on
- Schedule it to run each day
- Do more cleaning to make the data more useful
- Make the code a more readable