Welcome to our cutting-edge Tunisian Real Estate Data Pipeline! This project seamlessly integrates the power of Apache Kafka, Spark, Airflow, and Superset to process and visualize real estate data in real-time. By combining data from diverse sources, we offer a holistic view of the Tunisian property market, enabling data-driven decision-making for investors, agents, and policymakers alike. 🚀 Our pipeline is designed to:
- Aggregate data from multiple Tunisian real estate sources
- Process and analyze information in near real-time
- Generate comprehensive insights into market trends
- Provide interactive visualizations for intuitive data exploration
Whether you're tracking property prices, monitoring market fluctuations, or identifying emerging hotspots, our Tunisia Real Estate Data Pipeline is your go-to solution for staying ahead in the dynamic world of Tunisian real estate.
- 🌟 Introduction
- 🏗️ Architecture
- 📊 Visualization Dashboard
- 🛠️ Prerequisites
- 🚀 Installation
- 🔧 Project Setup
- 🌈 Environment Variables
- 💾 Data Persistence
- 🐛 Troubleshooting
- 🤝 Contributing
- 📄 License
Dive into the dynamic world of Tunisian real estate analytics with our comprehensive data pipeline! This project harnesses data from multiple prominent sources including Tecnocasa, Remxx, and other key players in the Tunisian real estate market. Our robust pipeline ensures real-time analytics and delivers actionable insights into the ever-evolving property landscape. From data ingestion to insightful visualizations, we've got you covered! 📊🏘️
Here's a high-level overview of our pipeline architecture:
Our interactive visualization dashboard, powered by Apache Superset, provides real-time insights into the Tunisian real estate market.
Before embarking on this data journey, make sure you have:
-
Git 🐙
-
Docker 🐳 (Install Docker)
-
Docker Compose 🐋 (For Linux users):
sudo curl -L "https://github.com/docker/compose/releases/download/v2.20.0/docker-compose-$(uname -s)-$(uname -m)" -o /usr/local/bin/docker-compose sudo chmod +x /usr/local/bin/docker-compose docker-compose --version
-
Clone this repository:
git clone https://github.com/yourusername/real-estate-pipeline.git cd real-estate-pipeline
-
Install Python requirements (for local development):
pip install -r requirements.txt
-
Initialize Airflow:
docker-compose up airflow-init
-
Launch the pipeline:
docker-compose up --build -d
-
Access the services:
- 🎨 Apache Superset: http://localhost:8088
- 🌬️ Airflow Web UI: http://localhost:8080
- 🎛️ Kafka Control Center: http://localhost:9021
-
Stop the pipeline:
docker-compose down
Customize your pipeline by tweaking the .env
file. It's like choosing the perfect paint color for your house! 🎨
We use Docker volumes to keep your data safe and sound, even when containers take a nap. 😴
If things go sideways, check the logs:
docker-compose logs -f <service_name>
For Airflow logs:
docker-compose logs -f airflow
Got ideas? We love them! Fork the repo, make your changes, and send us a pull request. Let's build something amazing together! 🤜🤛
This project is licensed under the MIT License. Check out the LICENSE file for the fine print.
Built with ❤️ by Melek Msakni