Dieses Stellenangebot ist nicht mehr verfügbar
Über
Strong hands-on expertise with:
PySpark
(RDD, Data Frames, Spark SQL, performance tuning) DPL
(Data Pipeline Language / relevant tool-specific DPL)
Proficiency in
Python
for data engineering workflows. Experience with distributed computing and big data technologies (Spark, Hadoop, Delta Lake). Strong SQL skills and experience with relational and NoSQL databases. Experience building ETL/ELT pipelines on cloud platforms (AWS / Azure / GCP). Familiarity with CI/CD, Git, and containerization (Docker/Kubernetes) is a plus. Bachelor's or Master's in Computer Science, Engineering, or related field. Preferred Skills
Experience with orchestration tools (Airflow, ADF, Argo, Prefect). Knowledge of data warehousing concepts (Star schema, SCD, normalization). Experience with streaming platforms (Kafka, Kinesis, Spark Streaming). Exposure to data governance, security, and compliance frameworks. Experience working in Agile environments.
Sprachkenntnisse
- English
Hinweis für Nutzer
Dieses Stellenangebot wurde von einem unserer Partner veröffentlicht. Sie können das Originalangebot einsehen hier.