Using Machine Learning and NLP techniques, infer what the customers are not buying for the supplier and target these gaps with personalized sales promotions. The product is used to assist Sales Manages in increasing basket sizes for their customers.
In stock exchange companies, the volume of collected data in- creases daily. Also market value for these data increases. Considering these dynamic features of the project scope, we created an architecture where historical data could be provided to users and allowing them to enrich this data in their own scope.
Project aims to create a decoupled architecture to process and persist data on daily basis, collected into GCP. Also, in the scope of this project, we defined the CI/CD strategy for all the components of the data pipeline.
Create a self service framework for ingesting data from different sources and preprocessing, keeping in mind requirements from end consumer of these data. The framework is designed to work in a cloud native environment. A set of HTTP Rest APIs was made available to users to provide necessary functionality around on-demand AWS EMR clusters.
Implemented a Big Data Platform for ingestion, storage, and processing of batch and real-time data from a variety of data sources within the customer. As part of that, a centralized search engine was designed and implemented for company-wide knowledge and documentation of data sources. In another project, we aimed to transfer the persistence layer of one analytic application from Oracle to Impala. Data were provided on regular basis, and the result of processing with Spark and Scala was a set of tables with more than 10000 columns. These jobs were scheduled on Apache Airflow.
Data and Information Management
Certifications:
Data Engineer or Machine Learning Engineer building production grade data pipeline or model training and deployment pipelines
WORK EXPERIENCE:
03/2021 – present
Rolle: Senior Consultant, ML/Data Engineer
Kunde: Freelance
Einsatzort: Munich, Germany
12/2016 – 03/2021
Rolle: Senior Consultant, ML/Data Engineer
Kunde: Data Reply
Einsatzort: Munich, Germany
EARLIER WORK EXPERIENCES:
01/2014 – 07/2015
Rolle: Student Helper
Kunde: RWTH Aachen University, Informatik Zentrum, Lehrstuhl für Informatik 5
Aufgaben:
Development of Web Services to allow storing and annotating multimedia data using relational and non-relational databases in microservice architecture.
09/2010 – 08/2013
Rolle: Software Developer
Kunde: Helius Systems, Tirana, Albania
Aufgaben:
Full-stack software developer in several projects used by up to 1500 users. Languages used: VB, SQL, C# and ASP.NET, Crystal Reports.
Cloud Providers:
I am a goal-oriented technology consultant, specialized in providing production-grade solutions to Data and Machine Learning related challenges.
The focus of my work is to design and develop scalable, extensible solutions that integrate well in your current organizational setup.
I'm experienced in the following topics:
- Cloud Development (AWS, GCP, Kubernetes)
- Data Processing (Spark, pandas, numpy)
- Machine Learning (scikit-learn, PyTorch, Tensorflow)
- Orchestration (Airflow, Kubeflow)
Facts about me:
- 10+ years of experience in Software Engineering
- 5 years of experience Data Engineering
- 4 years of experience in Public Cloud Providers (GCP, AWS)
- 5 years of experience in Python
- 3 years of experience in Scala
- 3 years of experience in Python
- Certified AWS Architect Professional
- Certified GCP Architect
Using Machine Learning and NLP techniques, infer what the customers are not buying for the supplier and target these gaps with personalized sales promotions. The product is used to assist Sales Manages in increasing basket sizes for their customers.
In stock exchange companies, the volume of collected data in- creases daily. Also market value for these data increases. Considering these dynamic features of the project scope, we created an architecture where historical data could be provided to users and allowing them to enrich this data in their own scope.
Project aims to create a decoupled architecture to process and persist data on daily basis, collected into GCP. Also, in the scope of this project, we defined the CI/CD strategy for all the components of the data pipeline.
Create a self service framework for ingesting data from different sources and preprocessing, keeping in mind requirements from end consumer of these data. The framework is designed to work in a cloud native environment. A set of HTTP Rest APIs was made available to users to provide necessary functionality around on-demand AWS EMR clusters.
Implemented a Big Data Platform for ingestion, storage, and processing of batch and real-time data from a variety of data sources within the customer. As part of that, a centralized search engine was designed and implemented for company-wide knowledge and documentation of data sources. In another project, we aimed to transfer the persistence layer of one analytic application from Oracle to Impala. Data were provided on regular basis, and the result of processing with Spark and Scala was a set of tables with more than 10000 columns. These jobs were scheduled on Apache Airflow.
Data and Information Management
Certifications:
Data Engineer or Machine Learning Engineer building production grade data pipeline or model training and deployment pipelines
WORK EXPERIENCE:
03/2021 – present
Rolle: Senior Consultant, ML/Data Engineer
Kunde: Freelance
Einsatzort: Munich, Germany
12/2016 – 03/2021
Rolle: Senior Consultant, ML/Data Engineer
Kunde: Data Reply
Einsatzort: Munich, Germany
EARLIER WORK EXPERIENCES:
01/2014 – 07/2015
Rolle: Student Helper
Kunde: RWTH Aachen University, Informatik Zentrum, Lehrstuhl für Informatik 5
Aufgaben:
Development of Web Services to allow storing and annotating multimedia data using relational and non-relational databases in microservice architecture.
09/2010 – 08/2013
Rolle: Software Developer
Kunde: Helius Systems, Tirana, Albania
Aufgaben:
Full-stack software developer in several projects used by up to 1500 users. Languages used: VB, SQL, C# and ASP.NET, Crystal Reports.
Cloud Providers:
I am a goal-oriented technology consultant, specialized in providing production-grade solutions to Data and Machine Learning related challenges.
The focus of my work is to design and develop scalable, extensible solutions that integrate well in your current organizational setup.
I'm experienced in the following topics:
- Cloud Development (AWS, GCP, Kubernetes)
- Data Processing (Spark, pandas, numpy)
- Machine Learning (scikit-learn, PyTorch, Tensorflow)
- Orchestration (Airflow, Kubeflow)
Facts about me:
- 10+ years of experience in Software Engineering
- 5 years of experience Data Engineering
- 4 years of experience in Public Cloud Providers (GCP, AWS)
- 5 years of experience in Python
- 3 years of experience in Scala
- 3 years of experience in Python
- Certified AWS Architect Professional
- Certified GCP Architect