Data is extracted from an Azure cloud SQL server using Azure Data Factory to Azure Data Lake Gen2 and saved as Parquet files. The data is then transformed within Databricks and then modelled into a star schema consisting of a fact table and multiple dimensions.
The final result is the following star schema:
ETL
Extracting the data from the SQL server to the Data Lake was done with Azure Data Factory.