Über
Build real-time and batch ingestion pipelines into Snowflake using Snowpipe, Kafka Connect, Fivetran, Matillion,
Informatica , or dbt.
Automate incremental data processing with Streams & Tasks to support CDC (Change Data Capture).
Use Zero-Copy Cloning for environment management, testing, and sandboxing.
Apply Time Travel and Fail-safe features for data recovery and auditing.
Develop data transformation logic in Snowpark for Python/SQL/Scala to push compute directly into Snowflake.
Design integrations with cloud storage (S3, Azure ADLS, GCS) for staging and external tables.
Implement data sharing and data marketplace solutions via Snowflake Secure Data Sharing and Snowflake Marketplace.
Enable semi-structured data handling (JSON, Avro, Parquet, ORC, XML) using VARIANT columns and lateral flattening.
Integrate Snowflake with BI tools (Power BI, Tableau) via live connections and semantic layers.
Implement RBAC (Role-Based Access Control), Row Access Policies, and Dynamic Data Masking for data security.
Optimize compute usage with multi-cluster warehouses, resource monitors, and query performance tuning.
Manage cost optimization strategies (warehouse auto-suspend, query profiling, storage/compute separation).
Integrate with data catalog & governance platforms (Collibra, Alation, Informatica CDGC) using Snowflake metadata and APIs.
Work with domain teams to deliver data products leveraging Snowflake’s data mesh-friendly features.
Collaborate with architects to design a Snowflake-centric data fabric integrated with ETL/ELT and API layers.
Support CI/CD automation for Snowflake code deployment using GitHub Actions, Azure DevOps, or dbt Cloud.
Qualifications Education : Bachelor’s or Master’s in Computer Science, Data Engineering, or related field. Experience: 10+ years of data engineering experience, with 5+ years in
Snowflake Data Cloud .
Expertise in
SQL
optimization and
Snowflake performance tuning .
Hands-on with
Snowpipe, Streams & Tasks, Snowpark, Zero-Copy Cloning, and Secure Data Sharing.
Proficiency in Python, Scala, or Java for Snowpark development.
Experience integrating with cloud platforms like
AWS .
Exposure to ETL/ELT tools ( Informatica , Matillion, Fivetran).
Familiarity with CI/CD, Git, DevOps practices for data operations.
Preferred Certifications:
SnowPro Core Certification
Key Skills Snowflake-native feature design and implementation (Snowpark, Streams, Time Travel, Secure Data Sharing)
Data ingestion (Snowpipe, CDC, Kafka, Fivetran)
Semi-structured data handling (VARIANT, JSON, Avro, Parquet)
Advanced SQL and performance tuning
Data governance (RBAC, masking, lineage, catalogs)
Cloud data platform integrations (AWS S3, Azure ADLS, GCP GCS)
BI and analytics tool integration
Cost optimization and warehouse orchestration
Sprachkenntnisse
- English
Hinweis für Nutzer
Dieses Stellenangebot stammt von einer Partnerplattform von TieTalent. Klicken Sie auf „Jetzt Bewerben“, um Ihre Bewerbung direkt auf deren Website einzureichen.