Cette offre d'emploi n'est plus disponible
Senior Data Engineer- Data Science & AI
Sedgwick
- United States
- United States
À propos
Key Responsibilities
Hybrid Data Pipeline Execution: Design and implement robust ETL/ELT pipelines to ingest data from legacy on‑prem sources, AWS (S3/RDS), and Azure (Blob/SQL), centralizing it for consumption in Snowflake and AI services.
Engineering for Data Science: Build and maintain Feature Stores and specialized datasets optimized for machine learning, ensuring Data Scientists have immediate access to clean, versioned, and statistically valid data.
Engineering for AI (RAG & LLMs): Develop the data pipelines required for Generative AI, including the automated extraction, chunking, and loading of unstructured data into vector stores across AWS and Azure.
Snowflake Power‑User Execution: Act as the technical lead for our Snowflake data warehouse, implementing sophisticated data modeling, Snowpipe automation, and compute optimization to support high‑concurrency AI workloads.
Legacy “Back‑Reach” Engineering: Execute non‑invasive data extraction patterns to unlock mission‑critical data from decades‑old on‑premise systems without disrupting core business operations.
Multi‑Cloud Orchestration: Manage complex, cross‑platform data workflows using Airflow, Step Functions, or Azure Data Factory, ensuring the synchronization of data across our multi‑cloud AI posture.
IT & Security Diplomacy: Partner directly with central IT, Database Administrators, and Security teams to solve connectivity hurdles (PrivateLink, IAM, firewalls) and secure “license to operate” for new data flows.
Data Quality for Model Integrity: Implement automated validation and observability layers to detect data drift and quality issues that could compromise the accuracy of production AI and Data Science models.
Cost & Performance Management: Drive the efficiency of our data stack by optimizing storage and query performance in Snowflake, AWS, and Azure to manage the ROI of the Transformation Office.
Direct Stakeholder Collaboration: Work as a dedicated engineering partner to MLOps and Data Science teams to rapidly iterate on data requirements for evolving AI use cases.
Qualifications
Education: Bachelor’s degree in Computer Science, Data Engineering, or a related field is required. A Master’s degree is highly desirable.
Proven Execution: 6+ years of hands‑on data engineering experience, with a track record of building production‑grade pipelines for Data Science and AI in multi‑cloud environments.
Snowflake Mastery: Expert‑level proficiency in Snowflake architecture, including data sharing, performance tuning, and the integration of Snowflake with external cloud AI services.
Multi‑Cloud Proficiency: Advanced, hands‑on knowledge of AWS (S3, Glue, Lambda) and Azure (Data Factory, Synapse) data services.
Technical Stack: Mastery of Python, SQL, and PySpark. Deep experience with data orchestration and containerization (Docker).
Legacy Expertise: Proven ability to interface with “old world” tech (on‑premise SQL, Mainframe extracts, flat files) and transform it for modern cloud consumption.
AI/DS Fluency: A strong understanding of the specific data needs for Machine Learning (feature engineering) and Generative AI (vectorization and embedding pipelines).
Execution Mindset: A “get‑it‑done” attitude, capable of navigating enterprise bureaucracy and technical debt to ship code at the speed required by a Transformation Office.
Sedgwick is an Equal Opportunity Employer and a Drug‑Free Workplace.
#J-18808-Ljbffr
Compétences linguistiques
- English
Avis aux utilisateurs
Cette offre a été publiée par l’un de nos partenaires. Vous pouvez consulter l’offre originale ici.