Über
Location - Dallas, TX (3 days per week in office)
Duration - Long Term
Required Skills:
Data Integration & ETL:
Build and manage scalable data pipelines to ingest data from diverse sources (ERP, CRM, APIs, S3, SQL databases) into Foundry. •
Ontology Modeling:
Define and maintain the "Ontology"-the platform's semantic layer-which maps technical data to real-world business objects (e.g., "Aircraft," "Customer," or "Invoice"). •
Pipeline Development:
Write and optimize data transformations using
PySpark ,
SQL , or
Java
within Foundry's Code Repositories. •
Application Building:
Develop front-end operational applications and interactive dashboards using low-code/pro-code tools like
Workshop
and
Slate . •
AIP Integration:
Implement Artificial Intelligence Platform (AIP) features, such as LLM-backed functions and agents, to automate workflows. •
Data Governance & Security:
Configure granular access controls, data health monitors, and lineage tracking to ensure compliance and reliability.
## Core Technical Skills •
Languages:
High proficiency in
Python (PySpark)
and
SQL
is mandatory. Knowledge of Java, TypeScript, or JavaScript is often required for front-end customization. •
Big Data:
Understanding of distributed computing (Spark), data warehousing concepts, and schema design (Star, Snowflake, etc.). •
DevOps:
Experience with Git-based version control, CI/CD practices, and debugging complex data workflows. •
Cloud Architecture:
Familiarity with AWS, Azure, or GCP environments where Foundry is typically hosted.
Sprachkenntnisse
- English
Hinweis für Nutzer
Dieses Stellenangebot stammt von einer Partnerplattform von TieTalent. Klicken Sie auf „Jetzt Bewerben“, um Ihre Bewerbung direkt auf deren Website einzureichen.