XX
Data EngineerAmeware GroupUnited States

This job offer is no longer available

XX

Data Engineer

Ameware Group
  • US
    United States
  • US
    United States

About

Client:
NDA (USA) Location:
remote (LATAM & Europe) Employment:
full-time (40 hours per week)
Our client is a well‑known US‑based rewards marketplace that helps users earn for engaging with digital offers, while delivering strong value to advertisers through performance‑driven solutions. We are looking for a skilled and detail‑oriented Data Engineer who will play a key role in building and maintaining scalable data infrastructure within the AWS ecosystem.
The ideal candidate has solid hands‑on experience with Apache Airflow, Snowflake, and Python, and knows how to design efficient, fault‑tolerant pipelines. This role is perfect for someone who enjoys solving complex data challenges, optimizing workflows, and collaborating across product and engineering teams.
Requirements
3–5 years of proven experience as a
Data Engineer , with strong expertise in designing and maintaining production‑grade data pipelines.
2+ years of experience with
Apache Airflow (MWAA) , including building and debugging complex DAGs, and applying best practices for error handling and idempotency.
2+ years of practical experience with
Snowflake , including advanced SQL (MERGE, COPY INTO), query tuning, and orchestration tasks.
3+ years of experience with
Python , including object‑oriented programming,
Pandas , and concurrent execution (e.g., ThreadPoolExecutor for parallel data processing).
3+ years of experience writing complex
SQL
queries (CTEs, window functions, stored procedures) and optimizing them for large datasets.
2+ years of experience working with
AWS
(S3, IAM, Lambda).
Experience with
Infrastructure as Code
tools (Terraform or CloudFormation) is a strong plus.
Strong understanding of
ETL/ELT principles , data modeling (star schema, slowly changing dimensions), and data quality management.
Experience troubleshooting production data issues, including debugging race conditions, performance bottlenecks, and data inconsistencies.
Excellent communication skills, attention to detail, and a proactive, team‑oriented mindset.
Nice to have
Experience with marketing and mobile attribution data sources (e.g., Adjust, Singular, AppsFlyer).
Familiarity with Looker or other BI visualization tools.
Experience with ETL/Reverse ETL platforms (Hevo, Hightouch, Census).
Knowledge of dbt or Great Expectations.
Certifications in Snowflake, AWS, or Airflow will be a plus.
Responsibilities
Design, develop, and maintain scalable ETL/ELT pipelines using Python, Apache Airflow (MWAA), and Snowflake.
Build data ingestion workflows from multiple sources, including AWS S3, Aurora MySQL, MongoDB, BigQuery, and third‑party APIs (Singular, AppsFlyer, Google Ads, Facebook Ads, TikTok Ads).
Implement processes for data validation, cleansing, and deduplication to ensure high accuracy and reliability.
Apply a declarative and idempotent approach to pipeline design to support reliable backfills and recovery.
Set up monitoring, logging, and alerting systems (e.g., Slack notifications) for data operations and pipeline performance tracking.
Collaborate with Data, Product, and Engineering teams to translate business needs into reliable technical solutions.
Continuously optimize data workflows and SQL queries for performance and cost efficiency.
Ensure adherence to data quality and governance standards across all processes.
We offer
Remote work in an international company in the USA.
Competitive salary in the USD.
Flexible working hours to help you manage your work/life balance.
Fully remote service/support
Career and professional growth.
Warm and friendly attitude to every specialist.
#J-18808-Ljbffr
  • United States

Languages

  • English
Notice for Users

This job was posted by one of our partners. You can view the original job source here.