XX
Senior Data EngineerCurotecUnited States
XX

Senior Data Engineer

Curotec
  • US
    United States
  • US
    United States

Über

divh2Senior Data Engineer/h2pThis is a remote position. We are seeking a Senior Data Engineer to support the ingestion, processing, and synchronization of data across our analytics platform. This role focuses on using Python Notebooks to ingest data via APIs into Microsoft Fabrics Data Lake and Data Warehouse, with some data being synced to a Synapse Analytics database for broader reporting needs. The ideal candidate will have hands-on experience working with API-based data ingestion and modern data architectures, including implementing Medallion layer architecture (Bronze, Silver, Gold) for optimal data organization and quality management, with bonus points for exposure to marketing APIs like Google Ads, Google Business Profile, and Google Analytics 4. This is a remote position. We welcome applicants globally, but this role has a preference for LATAM candidates to ensure smoother collaboration with our existing team./ppKey Responsibilities/pulliBuild and maintain Python Notebooks to ingest data from third-party APIs/liliDesign and implement Medallion layer architecture (Bronze, Silver, Gold) for structured data organization and progressive data refinement/liliStore and manage data within Microsoft Fabrics Data Lake and Warehouse using delta parquet file formats/liliSet up data pipelines and sync key datasets to Azure Synapse Analytics/liliDevelop PySpark-based data transformation processes across Bronze, Silver, and Gold layers/liliCollaborate with developers, analysts, and stakeholders to ensure data availability and accuracy/liliMonitor, test, and optimize data flows for reliability and performance/liliDocument processes and contribute to best practices for data ingestion and transformation/li/ulpTech Stack Youll Use/pulliIngestion Processing:/liulliPython (Notebooks)/liliPySpark/li/ulliStorage Warehousing:/liulliMicrosoft Fabric Data Lake Data Warehouse/liliDelta Parquet files/li/ulliSync Reporting:/liulliAzure Synapse Analytics/li/ulliCloud Tooling:/liulliAzure Data Factory, Azure DevOps/li/ul/ulpRequirements:/pulliStrong experience with Python for data ingestion and transformation/liliProficiency with PySpark for large-scale data processing/liliProficiency in working with RESTful APIs and handling large datasets/liliExperience with Microsoft Fabric or similar modern data platforms/liliUnderstanding of Medallion architecture (Bronze, Silver, Gold layers) and data lakehouse concepts/liliExperience working with Delta Lake and parquet file formats/liliUnderstanding of data warehousing concepts and performance tuning/liliFamiliarity with cloud-based workflows, especially within the Azure ecosystem/li/ulpNice to Have:/pulliExperience with marketing APIs such as Google Ads or Google Analytics 4/liliFamiliarity with Azure Synapse and Data Factory pipeline design/liliUnderstanding of data modeling for analytics and reporting use cases/liliExperience with AI coding tools/liliExperience with Fivetran, Aribyte, and Riverly/li/ul/div
  • United States

Sprachkenntnisse

  • English
Hinweis für Nutzer

Dieses Stellenangebot stammt von einer Partnerplattform von TieTalent. Klicken Sie auf „Jetzt Bewerben“, um Ihre Bewerbung direkt auf deren Website einzureichen.