À propos
Bentonville, AR (Onsite) Experience:
8+ Years Total | 4+ Years GCP (Recent) Open for W2 Only Role Overview We are seeking an experienced
Senior Data Engineer
with strong
Spark and Streaming
expertise to build real-time, scalable data pipelines. The ideal candidate will work with
Spark, Kafka, Scala/Python, and GCP
to ingest, transform, and deliver data for analytics and machine learning. Key Responsibilities
Design, develop, and maintain
ETL/ELT pipelines
for batch and real-time data processing. Build scalable data pipelines using
Spark (PySpark/Scala)
and streaming technologies (
Kafka/Flink ). Develop and optimize
data architectures
including:
Data Lakes Data Warehouses (BigQuery) Streaming platforms
Perform
performance tuning
of Spark jobs, SQL queries, and workflows for speed and cost efficiency. Implement
data quality frameworks , monitoring, and alerting to ensure accuracy and reliability. Required Skills & Qualifications Programming
Strong proficiency in
Python and SQL Experience with
Scala or Java
(preferred) Big Data & Streaming
Expertise in
Apache Spark
(Spark SQL, DataFrames, Spark Streaming) Experience with
Apache Kafka or Pub/Sub Cloud
Strong hands-on experience with
Google Cloud Platform (GCP) Exposure to
Azure Data Services
is a plus Data & Tools
Data Warehousing:
BigQuery, Snowflake, Redshift Orchestration & DevOps tools:
Airflow Databricks Docker Kubernetes
(nice to have)
Compétences linguistiques
- English
Avis aux utilisateurs
Cette offre provient d’une plateforme partenaire de TieTalent. Cliquez sur « Postuler maintenant » pour soumettre votre candidature directement sur leur site.