You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+32-8Lines changed: 32 additions & 8 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -94,22 +94,46 @@ Kedro-Energy-Forecasting/
94
94
│
95
95
├── .gitignore # Untracked files to ignore
96
96
├── Makefile # Set of tasks to be executed
97
+
├── Dockerfile # Instructions for building a Docker image
98
+
├── .dockerignore # Files and directories to ignore in Docker builds
97
99
├── README.md # Project documentation and setup guide
98
100
└── requirements.txt # Project dependencies
99
101
```
100
102
101
103
## 🚀 Getting Started
102
104
103
-
Turn**raw CSV data** into a **trained pickle Machine Learning model** with these steps:
105
+
First,**Clone the Repository** to download a copy of the code onto your local machine, and before diving into transforming **raw data** into a **trained pickle Machine Learning model**, please note:
104
106
105
-
1.**Clone the Repository**: Download a copy of the code to your computer.
106
-
2.**Set Up the Environment**: Create a virtual environment using Conda or venv.
107
-
3.**Install Dependencies**: Run `pip install -r requirements.txt` in your environment to install the required libraries.
108
-
4.**Run the Kedro Pipeline**: `make run` or `kedro run` – and witness magic 🪄
109
-
5.**Review the Results**: After running the pipeline, look in the `04_reporting` and `05_model_output` directories to see your model's performance and results.
110
-
6.**(Optional) Launch Kedro Viz**: To see a visual representation of your pipeline, run `make viz` or `kedro run viz`.
107
+
🔴 **Important Preparation Steps**:
108
+
- If you intend to run the code, it's better to remove the following directories if they exist: `data/02_processed`, `data/03_training_data`, `data/04_reporting`, and `data/05_model_output`. These directories will be regenerated or overwritten after executing the pipeline. They are **included** in the version control to **give you a preview of the expected outcomes**.
111
109
112
-
_Need guidance on commands? Peek into the **Makefile** or use `kedro --help` for assistance._
110
+
111
+
112
+
### Standard Method (Conda / venv) 🌿
113
+
114
+
Adopt this method if you prefer a traditional Python development environment setup using Conda or venv.
115
+
116
+
1.**Set Up the Environment**: Initialize a virtual environment with Conda or venv to isolate and manage your project's dependencies.
117
+
118
+
2.**Install Dependencies**: Inside your virtual environment, execute `pip install -r dev-requirements.txt` to install the necessary Python libraries.
119
+
120
+
3.**Run the Kedro Pipeline**: Trigger the pipeline processing by running `make run` or directly with `kedro run`. This step orchestrates your data transformation and modeling.
121
+
122
+
4.**Review the Results**: Inspect the `04_reporting` and `05_model_output` directories to assess the performance and outcomes of your models.
123
+
124
+
5.**(Optional) Explore with Kedro Viz**: To visually explore your pipeline's structure and data flows, initiate Kedro Viz using `make viz` or `kedro run viz`.
125
+
126
+
### Docker Method 🐳
127
+
128
+
Prefer this method for a containerized approach, ensuring a consistent development environment across different machines. Ensure Docker is operational on your system before you begin.
129
+
130
+
1.**Build the Docker Image**: Construct your Docker image with `make build` or `kedro docker build`. This command leverages `dev-requirements.txt` for environment setup. For advanced configurations, see the [Kedro Docker Plugin Documentation](https://github.com/kedro-org/kedro-plugins/tree/main/kedro-docker).
131
+
132
+
2.**Run the Pipeline Inside a Container**: Execute the pipeline within Docker using `make dockerun` or `kedro docker run`. Kedro-Docker meticulously handles volume mappings to ensure seamless data integration between your local setup and the Docker environment.
133
+
134
+
3.**Access the Results**: Upon completion, the `04_reporting` and `05_model_output` directories will contain your model's reports and trained files, ready for review.
135
+
136
+
For additional assistance or to explore more command options, refer to the **Makefile** or consult `kedro --help`.
0 commit comments