View job here
Your Key Responsibilities
- Design and implement large-scale data solutions using Big Data and Cloud technologies
- Design and develop conceptual data models for integrating, exploring and retrieving data from heterogeneous data sources and IT systems
- Implement and manage ETL processes, data transformations, data flows and service APIs
- Work with data virtualization, integration, and analytics technologies on our data platform
- Implement optimized data representations (views) and aggregations from multiple data sources
- Define and implement data ingestion, retention, lineage and access policies
- Define and implement data service API management and usage policies
- Define and implement data and data service catalog functionalities
- Work in interdisciplinary, cross-functional teams according to agile methodology
- Closely collaborate with business, digitalization and IT functions
Your Qualifications
- Successfully completed studies in Computer Science or related fields
- Proficient understanding of distributed and Cloud computing principles and technologies
- Experience with integration of data from multiple data sources and IT systems, e.g. ERP/SAP, DWH/SAP-BW, Web Services, CRM/Salesforce, MES, etc.
- Relevant knowledge in working with Cloud environments and services, preferably with Microsoft Azure, e.g. Azure Data Factory, Data Lake, IoT Hub, Databricks, etc.
- Strong programming skills in Java and/or Python; Software development experience is a plus
- Experience in data modeling (ER, UML) and proficiency in SQL
- Experience with relational databases such as Microsoft SQL Server and Oracle Postgres as well as experience with NoSQL databases such as MongoDB, Cassandra, InfluxDB, Stardog
- Knowledge of ETL techniques and frameworks, such as ADF, PowerCenter, NiFi, Sqoop
- Experience with Hadoop Ecosystem (Spark, HDFS, Hive, HBase), Containers (Docker, Kubernetes) and Distributed Messaging Platforms (Kafka, NATS.io) is a plus
- Fluent in English spoken and written, German is beneficial