À propos
Key Responsibilities Design and develop ETL/ELT workflows and data pipelines for batch and real-time processing. Build and maintain data pipelines for reporting and downstream applications using open-source frameworks and cloud technologies. Implement operational and analytical data stores leveraging Delta Lake and modern database concepts. Optimize data structures for performance and scalability across large datasets. Collaborate with architects and engineering teams to ensure alignment with target-state architecture. Apply best practices for data governance, lineage tracking, and metadata management, including integration with Google Dataplex for centralized governance and data quality enforcement. Develop, schedule, and orchestrate complex workflows using Apache Airflow, with strong proficiency in designing and managing Airflow DAGs. Troubleshoot and resolve issues in data pipelines and ensure high availability and reliability.
Compétences linguistiques
- English
Avis aux utilisateurs
Cette offre provient d’une plateforme partenaire de TieTalent. Cliquez sur « Postuler maintenant » pour soumettre votre candidature directement sur leur site.