Architected/Built a scalable data lake and analytics platform from scratch using AWS CDK, S3, AWS Glue, Apache Iceberg etc., to meet diverse business needs
Designed solutions for data replication from RDS MariaDB to S3 Data Lake and Amazon Redshift using AWS DMS, MSK (Amazon Managed Streaming for Apache Kafka) optimizing data availability for reporting and reducing performance bottlenecks on RDS
Prototyped and evaluated AWS services to align with business requirements, driving innovation in architecture and technology selection
Implemented real-time event ingestion pipelines with Amazon Kinesis Data Firehose, integrated with OpenSearch and S3 for advanced analytics
Developed secure, scalable RESTful APIs using API Gateway for seamless data access
Apache IcebergPythonSQLELTData ModellingAWS Database Migration Services (DMS)Amazon Managed Streaming for Apache Kafka (MSK)AWS Aurora MySQLAWS CDK (Typescript)GitLabKinesis Data FirehoseAmazon OpenSearch ServerlessAmazon API GatewayAmazon S3RedshiftAWS GlueLambdaDynamoDBRDS MariaDB etc.
Chargecloud GmbH, Köln (Germany)
1 year 3 months
2024-01 - now
various projects in the field of data engineering
FounderApache IcebergPythonSQL...
Founder
Database migration/development, data engineering consultation
Data migration coaching program
Apache IcebergPythonSQLELTData ModellingAWS Database Migration Services (DMS)Amazon Managed Streaming for Apache Kafka (MSK)AWS Aurora MySQLAWS CDK (Typescript)GitLabKinesis Data FirehoseAmazon OpenSearch ServerlessAmazon API GatewayAmazon S3RedshiftAWS GlueLambdaDynamoDBRDS MariaDB etc.
on request
Rastatt (Germany)
9 months
2023-04 - 2023-12
DataOps Helpdesk
Cloud Data EngineerPythonSQLETL...
Cloud Data Engineer
Provided pivotal support to internal customers on their Data Science and Engineering projects hosted on AWS, ensuring optimal data utilization
Developed and deployed a robust architecture designed to facilitate integration testing of all Terraform AWS modules of internal customers using Python and Boto3
Automated the testing process through GitLab, ensuring continuous testing of modules utilized by internal customers, thereby reducing errors, and enhancing reliability in daily operations
Developed robust data pipelines using Apache Airflow, transient EMR clusters, and GitLab, ensuring timely and accurate data delivery for the NLP project
Worked closely with data scientists to understand data requirements and ensured the provision of accurate data for model training and analysis
pearheaded the development and maintenance of ELT solutions
Business Intelligence EngineerPythonSQLELT...
Business Intelligence Engineer
Spearheaded the development and maintenance of ELT solutions on the Oracle platform, utilizing Oracle Data Integrator (ODI) to ensure seamless data extraction, loading, and transformation
Established Jenkins from scratch, creating a robust CI/CD data warehouse rollout pipeline, which streamlined the release of ELT packages and SQL scripts to various production and test instances
Enhanced deployment efficiency and reduced manual intervention by automating data warehouse releases
PythonSQLELTDatabase ModellingOracle Data Integrator (ODI)Oracle DatabaseJenkinsGroovy ScriptingPostgreSQL
Developed and maintained high-performance ETL solutions
Data EngineerPythonSQLETL...
Data Engineer
Developed and maintained high-performance ETL solutions utilizing Talend Open Studio (TOS) and MySQL, ensuring efficient data extraction, transformation, and loading to deliver data to the Business Intelligence/Data Science teams
Spearheaded the migration of legacy ETL scripts, originally written in PHP and Python, to Talend, ensuring a seamless transition with minimal disruption to ongoing operations
Developed real-time data pipelines using Apache Kafka to process daily web events (clicks, user interactions) for ingestion into MySQL, enabling web analytics capabilities
Stepped into the role of Interim Product Owner, effectively prioritizing and ranking backlogs on the sprint board, ensuring alignment with strategic objectives and timely delivery of data team projects
PythonSQLETLDatabase ModellingMySQL DatabaseRedisTalend Open StudioAWSKafkaMySQLPostgreSQL
Sovendus GmbH, Karlsruhe (Germany)
2 years
2016-12 - 2018-11
Designed and deployed high-performance data integration solutions
Database DeveloperPythonSQLETL...
Database Developer
Designed and deployed high-performance data integration solutions including Extraction, Transformation, and Loading (ETL) packages for Data Warehousing using Microsoft SQL Server Integration Services (SSIS)
Created tailored reports based on the varying needs of internal customers using Microsoft SQL Server Reporting Services (SSRS) and Tableau
Developed numerous web scraping programs with Python to extract data from internal sources, facilitating data availability for analysis and decision-making
PythonSQLETLDatabase ModellingSQL Server DatabaseSQL Server Integration Services (SSIS)SQL Server Reporting Services (SSRS)AWS Redshift & AuroraTableau
Misumi Europa GmbH, Frankfurt (Germany)
9 months
2016-03 - 2016-11
Designed a comprehensive process from scratch within SQL Server
Database Developer/Internship and Master ThesisSQLETLDatabase Modelling...
Database Developer/Internship and Master Thesis
Designed a comprehensive process from scratch within SQL Server to assess the quality of various data sources, ensuring robust data validation and quality assurance
Developed a front-end interface using Microsoft Access, providing users with an intuitive platform to interact with and manage data quality processes
Master Thesis Topic: on request
SQLETLDatabase ModellingSQL Server DatabaseSQL Server Integration Services (SSIS)Microsoft Access
Engaged in the project development phase, ensuring the quality and compatibility of code in adherence to Verizon?s Quality standards and industry best practices
Developed Teradata SQL test scripts, translating test procedures into executable scripts and executing them using Teradata SQL Assistant and BTEQ
Analyzed business and functional requirements, deriving detailed test plans, test cases, and procedures to ensure thorough testing of UNIX/Teradata-based Data Warehouse applications
Coordinated integration testing activities among various Data Warehouse teams and Upstream/Downstream application test teams, ensuring comprehensive testing and quality assurance
Acted as a pivotal Point of Contact (POC) among the Production, Development, and System Integration Testing (SIT) teams, ensuring seamless communication and issue tracking within the Production box
Conducted daily status calls with the development team and clients, ensuring all stakeholders were aligned and updated regarding project status and any emerging issues
Architected/Built a scalable data lake and analytics platform from scratch using AWS CDK, S3, AWS Glue, Apache Iceberg etc., to meet diverse business needs
Designed solutions for data replication from RDS MariaDB to S3 Data Lake and Amazon Redshift using AWS DMS, MSK (Amazon Managed Streaming for Apache Kafka) optimizing data availability for reporting and reducing performance bottlenecks on RDS
Prototyped and evaluated AWS services to align with business requirements, driving innovation in architecture and technology selection
Implemented real-time event ingestion pipelines with Amazon Kinesis Data Firehose, integrated with OpenSearch and S3 for advanced analytics
Developed secure, scalable RESTful APIs using API Gateway for seamless data access
Apache IcebergPythonSQLELTData ModellingAWS Database Migration Services (DMS)Amazon Managed Streaming for Apache Kafka (MSK)AWS Aurora MySQLAWS CDK (Typescript)GitLabKinesis Data FirehoseAmazon OpenSearch ServerlessAmazon API GatewayAmazon S3RedshiftAWS GlueLambdaDynamoDBRDS MariaDB etc.
Chargecloud GmbH, Köln (Germany)
1 year 3 months
2024-01 - now
various projects in the field of data engineering
FounderApache IcebergPythonSQL...
Founder
Database migration/development, data engineering consultation
Data migration coaching program
Apache IcebergPythonSQLELTData ModellingAWS Database Migration Services (DMS)Amazon Managed Streaming for Apache Kafka (MSK)AWS Aurora MySQLAWS CDK (Typescript)GitLabKinesis Data FirehoseAmazon OpenSearch ServerlessAmazon API GatewayAmazon S3RedshiftAWS GlueLambdaDynamoDBRDS MariaDB etc.
on request
Rastatt (Germany)
9 months
2023-04 - 2023-12
DataOps Helpdesk
Cloud Data EngineerPythonSQLETL...
Cloud Data Engineer
Provided pivotal support to internal customers on their Data Science and Engineering projects hosted on AWS, ensuring optimal data utilization
Developed and deployed a robust architecture designed to facilitate integration testing of all Terraform AWS modules of internal customers using Python and Boto3
Automated the testing process through GitLab, ensuring continuous testing of modules utilized by internal customers, thereby reducing errors, and enhancing reliability in daily operations
Developed robust data pipelines using Apache Airflow, transient EMR clusters, and GitLab, ensuring timely and accurate data delivery for the NLP project
Worked closely with data scientists to understand data requirements and ensured the provision of accurate data for model training and analysis
pearheaded the development and maintenance of ELT solutions
Business Intelligence EngineerPythonSQLELT...
Business Intelligence Engineer
Spearheaded the development and maintenance of ELT solutions on the Oracle platform, utilizing Oracle Data Integrator (ODI) to ensure seamless data extraction, loading, and transformation
Established Jenkins from scratch, creating a robust CI/CD data warehouse rollout pipeline, which streamlined the release of ELT packages and SQL scripts to various production and test instances
Enhanced deployment efficiency and reduced manual intervention by automating data warehouse releases
PythonSQLELTDatabase ModellingOracle Data Integrator (ODI)Oracle DatabaseJenkinsGroovy ScriptingPostgreSQL
Developed and maintained high-performance ETL solutions
Data EngineerPythonSQLETL...
Data Engineer
Developed and maintained high-performance ETL solutions utilizing Talend Open Studio (TOS) and MySQL, ensuring efficient data extraction, transformation, and loading to deliver data to the Business Intelligence/Data Science teams
Spearheaded the migration of legacy ETL scripts, originally written in PHP and Python, to Talend, ensuring a seamless transition with minimal disruption to ongoing operations
Developed real-time data pipelines using Apache Kafka to process daily web events (clicks, user interactions) for ingestion into MySQL, enabling web analytics capabilities
Stepped into the role of Interim Product Owner, effectively prioritizing and ranking backlogs on the sprint board, ensuring alignment with strategic objectives and timely delivery of data team projects
PythonSQLETLDatabase ModellingMySQL DatabaseRedisTalend Open StudioAWSKafkaMySQLPostgreSQL
Sovendus GmbH, Karlsruhe (Germany)
2 years
2016-12 - 2018-11
Designed and deployed high-performance data integration solutions
Database DeveloperPythonSQLETL...
Database Developer
Designed and deployed high-performance data integration solutions including Extraction, Transformation, and Loading (ETL) packages for Data Warehousing using Microsoft SQL Server Integration Services (SSIS)
Created tailored reports based on the varying needs of internal customers using Microsoft SQL Server Reporting Services (SSRS) and Tableau
Developed numerous web scraping programs with Python to extract data from internal sources, facilitating data availability for analysis and decision-making
PythonSQLETLDatabase ModellingSQL Server DatabaseSQL Server Integration Services (SSIS)SQL Server Reporting Services (SSRS)AWS Redshift & AuroraTableau
Misumi Europa GmbH, Frankfurt (Germany)
9 months
2016-03 - 2016-11
Designed a comprehensive process from scratch within SQL Server
Database Developer/Internship and Master ThesisSQLETLDatabase Modelling...
Database Developer/Internship and Master Thesis
Designed a comprehensive process from scratch within SQL Server to assess the quality of various data sources, ensuring robust data validation and quality assurance
Developed a front-end interface using Microsoft Access, providing users with an intuitive platform to interact with and manage data quality processes
Master Thesis Topic: on request
SQLETLDatabase ModellingSQL Server DatabaseSQL Server Integration Services (SSIS)Microsoft Access
Engaged in the project development phase, ensuring the quality and compatibility of code in adherence to Verizon?s Quality standards and industry best practices
Developed Teradata SQL test scripts, translating test procedures into executable scripts and executing them using Teradata SQL Assistant and BTEQ
Analyzed business and functional requirements, deriving detailed test plans, test cases, and procedures to ensure thorough testing of UNIX/Teradata-based Data Warehouse applications
Coordinated integration testing activities among various Data Warehouse teams and Upstream/Downstream application test teams, ensuring comprehensive testing and quality assurance
Acted as a pivotal Point of Contact (POC) among the Production, Development, and System Integration Testing (SIT) teams, ensuring seamless communication and issue tracking within the Production box
Conducted daily status calls with the development team and clients, ensuring all stakeholders were aligned and updated regarding project status and any emerging issues