I built a pipeline class that handles data extraction/fetch, transformation, displaying dashboard for n different pipelines and displays them separately on a dashboard.
Currently two pipelines are included, I can include as many more as I like.
- I use Open Meteo's weather data to predict if rain falls on the next day or not.
- I use Polygon's stock market data to predict the Closing price of a stock for the next day.
- Plotly Dash
- Plotly Express,
- Mysql
- Pandas,
- LSTM,
- XGBClassifier
- Datalake
- Data Warehouse
- ETL Functions
- Pipeline workflow.
- Config:
python -m venv .localenv
.localenv\Scripts\python -m pip install -r requirements.txt
- Start the workflow:
.localenv\Scripts\python -m src.start newopenmeteo
.localenv\Scripts\python -m src.start newpolygon
You can type newopenmeteo to fetch new data. or oldopenmeteo to get predict using data alreay in the data warehouse.
3. Start the dashboard:
.localenv\Scripts\python -m src.start dashboard
- Start multiple at a time:
.localenv\Scripts\python -m src.start newopenmeteo newpolygon dashboard
Iather data: https://open-meteo.com/en/docs
Stock market data: https://polygon.io/
They need to be stored in the .env file in:
project
├─keys
└─.env
POLYGON_KEY=
DB_USER=
DB_PASS=
DB_HOST=