Iβm an aspiring Data Engineer with hands-on experience building batch and real-time data pipelines using Python, SQL, Apache Airflow, Kafka, and Flink frm my learning and projects. I enjoy working on data ingestion, transformation, orchestration, and turning raw data into reliable, analytics-ready systems.
- Build ETL / ELT pipelines for batch and streaming data
- Work with Apache Kafka, Flink, Airflow, and PySpark
- Design and query relational and time-series databases
- Deploy and containerize data workflows using Docker
- Create real-time and analytical dashboards with Grafana
Programming & Querying:
Python, SQL
Data Engineering & Streaming:
Apache Kafka, Apache Flink, PySpark,pandas , numpy
Databases & Storage:
PostgreSQL, MySQL, Snowflake, TimescaleDB, Azure Data Lake Storage
Workflow Orchestration:
Apache Airflow
Cloud & DevOps:
AWS, Azure, Docker, Linux
Visualization & Monitoring:
Grafana
prometheus
Version Control:
Git, GitHub
- Real-time ingestion of live cryptocurrency prices from the Coinbase WebSocket API
- Event streaming with Apache Kafka (Confluent Cloud)
- Stream processing using Flink SQL with Avro schema validation
- Time-series storage in TimescaleDB (PostgreSQL)
- Sub-second latency dashboards built with Grafana
π Repository: (link here)
- Batch ingestion of weekly Formula One race data from a public API
- Data cleaning and transformation using Python & Pandas
- Workflow orchestration and scheduling with Apache Airflow
- Analytical storage in PostgreSQL and visualization in Grafana
- Fully containerized using Docker
π Repository: (link here)
- π§ Email: maxwelayal956@gmail.com
- πΌ LinkedIn: (Link here)
- π Portfolio: https://www.maxwel-ayal.tech
β‘ Currenlty Open to data engineering roles π



