High-Volume Data Batch Processing System

Jakub Dabkowski

Database Engineer
Data Engineer
Software Architect
Apache Airflow
Python
Snowflake

Project Overview

As a Senior Data Engineer, I led the development of a high-volume data batch processing system for a premier U.S. retail company. This system managed millions of data events daily, ensuring seamless operation and service efficiency for the food delivery platform.

Technologies Utilized:

Languages & Tools: Python, Apache Airflow, Snowflake, SQL, NoSQL, Docker, Bash, GitHub, ETL, API, Swagger, Grafana, Cassandrar

Project Details

A complex batch processing system for a premier U.S. retail company. This system was designed to handle and process millions of data events daily. It was crucial in managing vast volumes of transactional and customer data, ensuring seamless operation and service efficiency for the food delivery platform.

Accomplishments

Leadership: Led the development and optimization of the batch processing system, which is crucial for handling extensive data workloads.

Optimization: Acted as a technical mentor, guiding the team in complex data processing tasks and fostering skill development.

Automation: Utilized Airflow and Python to manage and automate the processing of hundreds of millions of events each day.

Scalability: Implemented Snowflake for data warehousing, enhancing the system’s scalability and performance.

Growth: Played a vital role in the technical growth of the team, emphasizing skill enhancement in data processing and system optimization.

Partner With Jakub
View Services

More Projects by Jakub