AI-Ready Data Platform Architect & Lead Data Engineer, Data/Information Architecture & Governance
Aktualisiert am 30.10.2025
Profil
Freiberufler / Selbstständiger
Remote-Arbeit
Verfügbar ab: 01.12.2025
Verfügbar zu: 100%
davon vor Ort: 5%
Data Engineering
Information Architecture
Conversational & Semantic Access to Enterprise Data
Data Architecture
Apache Spark
Apache Airflow
Azure Synapse
Microsoft Fabric
Data Lake / Lakehouse
ETL / Data Pipelines
SQL
Data Modeling
Data Integration
Data Governance
Metadata Management
Semantic Layer
LLM Integration
Data Intelligence Platform
Master Data Management
Big Data Engineer
Data Platform Architect
German
native
English
fluent
Spanish
B1 ? intermediate

Einsatzorte

Einsatzorte

Deutschland, Schweiz, Österreich
möglich

Projekte

Projekte

1 Jahr 10 Monate
2023-02 - 2024-11

Data Minions

Lead Data Engineer / Solution Designer Python Airflow Kubernetes ...
Lead Data Engineer / Solution Designer
Design and lead development of an internal data ingestion framework for the A1 Data Platform (Cloudera, Azure Synapse, Teradata DWH). This strategic tool replaced the existing ETL/ELT processes and handles loading for both the data lake and the DWH, ensuring unified and scalable data integration. The core principle: ?Easy data staging by specification, not implementation.? All required software components are provided within the framework ? the data engineer or scientist only specifies the source, basic transformations, and target platform. Execution is performed automatically by the framework via specific Airflow operators.
  • Architecture and design ownership for the framework
  • Lead development with Python, Apache Airflow, and Kubernetes
  • Migration and decommissioning of existing ETL processes
  • Coaching and technical support for other developers


Key Skills & Expertise:

  • Big Data Engineering (Hadoop, Spark, Kafka)
  • Data Governance & Data Ingestion
  • Leadership & Coaching

Python Airflow Kubernetes Data Engineer Cloudera Data Platform Azure Data Platform Hadoop Spark Kafka
A1 Telekom Austria
Wien
6 Jahre 11 Monate
2018-01 - 2024-11

Design and implementation of big data solutions

Lead Data Engineer Hadoop Spark Kafka ...
Lead Data Engineer
Design and implementation of big data solutions in the Hadoop ecosystem for both batch processing and near real-time use cases. Knowledge transfer and quality assurance as tutor and coach for the Data Engineering group.
  • Development of data pipelines with HDFS, Spark, Kafka, Hive, and Impala
  • Building scalable, fault-tolerant data processes
  • Implementation of data analytics use cases in close collaboration with business units
  • Development and rollout of pragmatic design guidelines and frameworks for the entire engineering team
  • Onboarding, tutoring, and coaching of new team members
  • Quality assurance in the Data Engineering group through solution design workshops, code reviews, and knowledge transfer in regular team meetings.

Key Skills & Expertise:
  • Hadoop ecosystem (HDFS, Spark, Hive, Kafka, Impala)
  • Data pipeline design & implementation
  • Near real-time data processing
  • Collaboration with business units
  • Team leadership & coaching
  • Quality management
Hadoop Spark Kafka Cloudera Data Platform Big Data HDFS Hive Impala
A1 Telekom Austria
Wien
5 Jahre 8 Monate
2019-01 - 2024-08

Data Lake Platform Migrations

Microsoft Azure Synapse Cloudera Data Platform Big Data Platforms
Repeated need to migrate the A1 Data Lake: Hortonworks >> Cloudera >> Cloudera CDP >> Microsoft Azure Synapse.
  • Leading role in planning and executing the migrations
  • Ensuring data quality and business continuity
  • Adapting data pipelines (batch and streaming) to new platform architectures
  • Knowledge transfer and training for the teams


Key Skills & Expertise:

  • Migration of big data platforms
  • Cloudera, Cloudera CDP, Azure Synapse
  • Cloud data architecture & data engineering
  • Project management & change management
Microsoft Azure Synapse Cloudera Data Platform Big Data Platforms
A1 Telekom Austria
1 Jahr
2022-01 - 2022-12

Netscout Data Integration

Lead Big Data Engineer Kubernetes OpenShift Apache Spark ...
Lead Big Data Engineer
Near real-time integration of 3.5 billion Location Session Records (LRS) per day from the network management system Netscout into the A1 Data Lake. The data includes geolocation and network performance metrics and forms the foundation for numerous analytical use cases.
  • Development of large-scale data extraction and integration
  • Performance optimization for high-volume data streams
  • Use of modern platforms (Kubernetes/OpenShift, HDFS, Spark, Linux shell)
  • Collaboration with analytics teams for downstream data processing

Key Skills & Expertise:
  • High-volume data engineering & streaming
  • Kubernetes, OpenShift, Apache Spark
  • Near real-time data integration
  • Use case enablement for analytics
Kubernetes OpenShift Apache Spark Kafka Bash
A1 Telekom Austria
Wien
3 Jahre 4 Monate
2014-02 - 2017-05

Enterprise Information Architecture

Information Architect Informationsarchitektur TM Forum Data Goverance ...
Information Architect

Scope & Focus Areas as a member of the Enterprise Architecture team:

  • Definition, implementation, and ongoing evolution of the Information Target Architecture.
  • Close integration of information architecture with other architecture disciplines (business [processes, capabilities], applications, technical, integration).
  • Identification and structuring of company-wide data assets.
  • Establishing structures for the transformation to the ?Data-Driven A1 Company.?


Tasks & Responsibilities:

  • Definition of target architecture and guiding principles for information architecture
  • Integration with business, process, and application architecture
  • Development and evolution of the corporate data model
  • Identification of business objects and central data domains


Key Skills & Expertise:

  • Enterprise architecture frameworks (TOGAF, TM Forum)
  • Information architecture & data governance
  • Data modeling & data management
  • Communication with business and IT architecture teams


Informationsarchitektur TM Forum Data Goverance Data Modeling Data Management Master Data Management DAMA DMBOK
A1 Telekom Austria
Wien

Position

Position



Kompetenzen

Kompetenzen

Top-Skills

Data Engineering Information Architecture Conversational & Semantic Access to Enterprise Data Data Architecture Apache Spark Apache Airflow Azure Synapse Microsoft Fabric Data Lake / Lakehouse ETL / Data Pipelines SQL Data Modeling Data Integration Data Governance Metadata Management Semantic Layer LLM Integration Data Intelligence Platform Master Data Management Big Data Engineer Data Platform Architect

Schwerpunkte

Information Architecture
Experte
Data Engineering
Experte

Produkte / Standards / Erfahrungen / Methoden

Apache Airflow
Experte
Apache Hadoop
Experte
Apache Kafka
Fortgeschritten
Apache Spark
Fortgeschritten
Big Data
Experte
Data Governance
Experte
Data Modeling
Experte
Information Architecture
Experte
ETL
Experte
Kubernetes / Openshift
Fortgeschritten
Microsoft Azure Data Platform
Fortgeschritten
Microsoft Fabric Data Platform
Fortgeschritten
Agile (Scrum)
Fortgeschritten
Data Lake / Lakehouse
Experte
Metadata Management
Experte
Semantic Layer
Fortgeschritten

KEY SKILLS

  • Agile Methodologies
  • Apache Airflow
  • Apache Hadoop
  • Apache Kafka
  • Apache Spark
  • Big Data
  • Cloudera Data Platform
  • Data Governance
  • Databases
  • Data Modeling
  • DevOps
  • Enterprise Architecture
  • Enterprise Information Architecture
  • Extract, Transform, Load (ETL)
  • Java
  • Kubernetes
  • Linux & Shell Programming
  • Microsoft Azure Data Platform
  • Microsoft Fabric Data Platform
  • Openshift
  • PySpark
  • Python
  • SQL
  • Scala
  • Software Design
  • Software development (incl. OO)
  • Telecommunication


Emerging Skills & Current Learning

  • In addition to my established expertise, I am actively expanding my skill set to integrate the latest technologies and methods into my Data Engineering practice:
    • Generative AI Applications ? exploring the use of GenAI to enhance and accelerate Data Engineering workflows.
    • Databricks Data Intelligence Platform ? gaining hands-on experience with Databricks for scalable data processing and analytics.
    • Data Science & Machine Learning ? currently building solid foundations through structured learning and practical work on a partner project (pro bono), applying ML basics in real-world contexts.
  • These ongoing learning paths are already enriching my core capabilities in Data Engineering, enabling me to design solutions that bridge modern AI methods, advanced data platforms, and established enterprise data practices.


EXPERTISE

Beyond the Basics

  • Enterprise Architektur Frameworks (TOGAF, TM-Forum)
  • European Union GDPR
  • Information Security & Data Privacy Management
  • Pseudonymization and anonymization procedures (k-anonymity, etc.)
  • Object-oriented analysis and conception
  • Project Management
  • Employee Management and Leadership


Selected Tools & Frameworks

  • Automic Automation Engine (formerly UC4)
  • Data Modeling with CA Erwin / Sysbase Powerdesigner
  • ETL: Teradata, AbInitio, Decision Stream, Sagent ETL
  • Implementation of Planning Systems
  • Power BI, QlikView, MS Reporting Services, IBM/Cognos Reporting
  • Jira & Confluence
  • MS Office incl. Programming
  • WordPress/Joomla/Typo3 CMS

Programmiersprachen

Python
Pyspark
Java
Scala
SQL
T-SQL, TD-SQL, PL/SQL
MDX
Shell scripting
C++
C#.Net
C
Visual Basic
VBA
VB.Net


Datenbanken

Microsoft Azure Synapse
Microsoft Fabric Lakehouse
Hadoop Ecosystem
Hive
Impala
Teradata RDBMS
Microsoft SQL Server
Oracle
PostgreSQL
Microsoft Analysis Services
mySQL
MS Access
MSM
Caché

Einsatzorte

Einsatzorte

Deutschland, Schweiz, Österreich
möglich

Projekte

Projekte

1 Jahr 10 Monate
2023-02 - 2024-11

Data Minions

Lead Data Engineer / Solution Designer Python Airflow Kubernetes ...
Lead Data Engineer / Solution Designer
Design and lead development of an internal data ingestion framework for the A1 Data Platform (Cloudera, Azure Synapse, Teradata DWH). This strategic tool replaced the existing ETL/ELT processes and handles loading for both the data lake and the DWH, ensuring unified and scalable data integration. The core principle: ?Easy data staging by specification, not implementation.? All required software components are provided within the framework ? the data engineer or scientist only specifies the source, basic transformations, and target platform. Execution is performed automatically by the framework via specific Airflow operators.
  • Architecture and design ownership for the framework
  • Lead development with Python, Apache Airflow, and Kubernetes
  • Migration and decommissioning of existing ETL processes
  • Coaching and technical support for other developers


Key Skills & Expertise:

  • Big Data Engineering (Hadoop, Spark, Kafka)
  • Data Governance & Data Ingestion
  • Leadership & Coaching

Python Airflow Kubernetes Data Engineer Cloudera Data Platform Azure Data Platform Hadoop Spark Kafka
A1 Telekom Austria
Wien
6 Jahre 11 Monate
2018-01 - 2024-11

Design and implementation of big data solutions

Lead Data Engineer Hadoop Spark Kafka ...
Lead Data Engineer
Design and implementation of big data solutions in the Hadoop ecosystem for both batch processing and near real-time use cases. Knowledge transfer and quality assurance as tutor and coach for the Data Engineering group.
  • Development of data pipelines with HDFS, Spark, Kafka, Hive, and Impala
  • Building scalable, fault-tolerant data processes
  • Implementation of data analytics use cases in close collaboration with business units
  • Development and rollout of pragmatic design guidelines and frameworks for the entire engineering team
  • Onboarding, tutoring, and coaching of new team members
  • Quality assurance in the Data Engineering group through solution design workshops, code reviews, and knowledge transfer in regular team meetings.

Key Skills & Expertise:
  • Hadoop ecosystem (HDFS, Spark, Hive, Kafka, Impala)
  • Data pipeline design & implementation
  • Near real-time data processing
  • Collaboration with business units
  • Team leadership & coaching
  • Quality management
Hadoop Spark Kafka Cloudera Data Platform Big Data HDFS Hive Impala
A1 Telekom Austria
Wien
5 Jahre 8 Monate
2019-01 - 2024-08

Data Lake Platform Migrations

Microsoft Azure Synapse Cloudera Data Platform Big Data Platforms
Repeated need to migrate the A1 Data Lake: Hortonworks >> Cloudera >> Cloudera CDP >> Microsoft Azure Synapse.
  • Leading role in planning and executing the migrations
  • Ensuring data quality and business continuity
  • Adapting data pipelines (batch and streaming) to new platform architectures
  • Knowledge transfer and training for the teams


Key Skills & Expertise:

  • Migration of big data platforms
  • Cloudera, Cloudera CDP, Azure Synapse
  • Cloud data architecture & data engineering
  • Project management & change management
Microsoft Azure Synapse Cloudera Data Platform Big Data Platforms
A1 Telekom Austria
1 Jahr
2022-01 - 2022-12

Netscout Data Integration

Lead Big Data Engineer Kubernetes OpenShift Apache Spark ...
Lead Big Data Engineer
Near real-time integration of 3.5 billion Location Session Records (LRS) per day from the network management system Netscout into the A1 Data Lake. The data includes geolocation and network performance metrics and forms the foundation for numerous analytical use cases.
  • Development of large-scale data extraction and integration
  • Performance optimization for high-volume data streams
  • Use of modern platforms (Kubernetes/OpenShift, HDFS, Spark, Linux shell)
  • Collaboration with analytics teams for downstream data processing

Key Skills & Expertise:
  • High-volume data engineering & streaming
  • Kubernetes, OpenShift, Apache Spark
  • Near real-time data integration
  • Use case enablement for analytics
Kubernetes OpenShift Apache Spark Kafka Bash
A1 Telekom Austria
Wien
3 Jahre 4 Monate
2014-02 - 2017-05

Enterprise Information Architecture

Information Architect Informationsarchitektur TM Forum Data Goverance ...
Information Architect

Scope & Focus Areas as a member of the Enterprise Architecture team:

  • Definition, implementation, and ongoing evolution of the Information Target Architecture.
  • Close integration of information architecture with other architecture disciplines (business [processes, capabilities], applications, technical, integration).
  • Identification and structuring of company-wide data assets.
  • Establishing structures for the transformation to the ?Data-Driven A1 Company.?


Tasks & Responsibilities:

  • Definition of target architecture and guiding principles for information architecture
  • Integration with business, process, and application architecture
  • Development and evolution of the corporate data model
  • Identification of business objects and central data domains


Key Skills & Expertise:

  • Enterprise architecture frameworks (TOGAF, TM Forum)
  • Information architecture & data governance
  • Data modeling & data management
  • Communication with business and IT architecture teams


Informationsarchitektur TM Forum Data Goverance Data Modeling Data Management Master Data Management DAMA DMBOK
A1 Telekom Austria
Wien

Position

Position



Kompetenzen

Kompetenzen

Top-Skills

Data Engineering Information Architecture Conversational & Semantic Access to Enterprise Data Data Architecture Apache Spark Apache Airflow Azure Synapse Microsoft Fabric Data Lake / Lakehouse ETL / Data Pipelines SQL Data Modeling Data Integration Data Governance Metadata Management Semantic Layer LLM Integration Data Intelligence Platform Master Data Management Big Data Engineer Data Platform Architect

Schwerpunkte

Information Architecture
Experte
Data Engineering
Experte

Produkte / Standards / Erfahrungen / Methoden

Apache Airflow
Experte
Apache Hadoop
Experte
Apache Kafka
Fortgeschritten
Apache Spark
Fortgeschritten
Big Data
Experte
Data Governance
Experte
Data Modeling
Experte
Information Architecture
Experte
ETL
Experte
Kubernetes / Openshift
Fortgeschritten
Microsoft Azure Data Platform
Fortgeschritten
Microsoft Fabric Data Platform
Fortgeschritten
Agile (Scrum)
Fortgeschritten
Data Lake / Lakehouse
Experte
Metadata Management
Experte
Semantic Layer
Fortgeschritten

KEY SKILLS

  • Agile Methodologies
  • Apache Airflow
  • Apache Hadoop
  • Apache Kafka
  • Apache Spark
  • Big Data
  • Cloudera Data Platform
  • Data Governance
  • Databases
  • Data Modeling
  • DevOps
  • Enterprise Architecture
  • Enterprise Information Architecture
  • Extract, Transform, Load (ETL)
  • Java
  • Kubernetes
  • Linux & Shell Programming
  • Microsoft Azure Data Platform
  • Microsoft Fabric Data Platform
  • Openshift
  • PySpark
  • Python
  • SQL
  • Scala
  • Software Design
  • Software development (incl. OO)
  • Telecommunication


Emerging Skills & Current Learning

  • In addition to my established expertise, I am actively expanding my skill set to integrate the latest technologies and methods into my Data Engineering practice:
    • Generative AI Applications ? exploring the use of GenAI to enhance and accelerate Data Engineering workflows.
    • Databricks Data Intelligence Platform ? gaining hands-on experience with Databricks for scalable data processing and analytics.
    • Data Science & Machine Learning ? currently building solid foundations through structured learning and practical work on a partner project (pro bono), applying ML basics in real-world contexts.
  • These ongoing learning paths are already enriching my core capabilities in Data Engineering, enabling me to design solutions that bridge modern AI methods, advanced data platforms, and established enterprise data practices.


EXPERTISE

Beyond the Basics

  • Enterprise Architektur Frameworks (TOGAF, TM-Forum)
  • European Union GDPR
  • Information Security & Data Privacy Management
  • Pseudonymization and anonymization procedures (k-anonymity, etc.)
  • Object-oriented analysis and conception
  • Project Management
  • Employee Management and Leadership


Selected Tools & Frameworks

  • Automic Automation Engine (formerly UC4)
  • Data Modeling with CA Erwin / Sysbase Powerdesigner
  • ETL: Teradata, AbInitio, Decision Stream, Sagent ETL
  • Implementation of Planning Systems
  • Power BI, QlikView, MS Reporting Services, IBM/Cognos Reporting
  • Jira & Confluence
  • MS Office incl. Programming
  • WordPress/Joomla/Typo3 CMS

Programmiersprachen

Python
Pyspark
Java
Scala
SQL
T-SQL, TD-SQL, PL/SQL
MDX
Shell scripting
C++
C#.Net
C
Visual Basic
VBA
VB.Net


Datenbanken

Microsoft Azure Synapse
Microsoft Fabric Lakehouse
Hadoop Ecosystem
Hive
Impala
Teradata RDBMS
Microsoft SQL Server
Oracle
PostgreSQL
Microsoft Analysis Services
mySQL
MS Access
MSM
Caché

Vertrauen Sie auf Randstad

Im Bereich Freelancing
Im Bereich Arbeitnehmerüberlassung / Personalvermittlung

Fragen?

Rufen Sie uns an +49 89 500316-300 oder schreiben Sie uns:

Das Freelancer-Portal

Direktester geht's nicht! Ganz einfach Freelancer finden und direkt Kontakt aufnehmen.