This project involves processing and transforming data using PySpark. The data is read from a Parquet file, transformed using window functions, and then saved as a table in Parquet format.
solution/run.py - main program solution/DataProcessor.py - class for data processing solution/quick_and_dirty - databricks notebook for quick solution
tests - folder with tests