Über
Location: Toronto, ON
Employment Type: Full-Time
About Cogency
Cogency is a consulting and technology services firm delivering enterprise digital transformation solutions across financial services and regulated industries. We specialize in building scalable digital platforms, automating business processes, and enabling data-driven decision-making through modern cloud and low-code technologies.
Position Summary:
We are seeking an experienced Data Scientist specializing in AI/ML and Generative AI to design, develop, deploy, and operationalize advanced machine learning solutions. This role requires deep expertise in model development, performance optimization, and production deployment using modern MLOps and LLMOps frameworks within cloud-based environments.
The ideal candidate combines strong research and modeling capabilities with hands-on experience delivering scalable, enterprise-grade AI systems leveraging AWS, Snowflake, and streaming technologies such as Kafka.
Key Responsibilities:
AI/ML & Generative AI Development
Design, develop, and deploy machine learning and generative AI models to solve complex business problems.
Build deep learning and large language model (LLM)-based solutions using TensorFlow and PyTorch.
Develop custom training pipelines, feature engineering processes, and model evaluation frameworks.
Fine-tune, optimize, and adapt pre-trained models for domain-specific use cases.
Model Optimization & Performance Management
Establish model performance benchmarks and evaluation metrics (accuracy, precision, recall, F1, AUC, latency, throughput, etc.).
Conduct hyperparameter tuning and model optimization to improve scalability and efficiency.
Monitor model drift, bias, and degradation in production environments.
Implement automated retraining and performance monitoring strategies.
MLOps & LLMOps Implementation
Design and implement scalable ML pipelines using MLOps best practices, including CI/CD for models.
Manage model versioning, artifact tracking, and reproducibility.
Automate model deployment workflows and rollback mechanisms.
Apply LLMOps principles for managing large language model lifecycle, prompt management, evaluation, and governance.
Ensure security, compliance, and governance standards for AI systems.
Cloud & Data Platform Integration
Deploy and manage ML solutions within AWS cloud environments.
Integrate models with Snowflake and other enterprise data platforms for scalable data access and inference.
Collaborate closely with data engineering teams to ensure efficient data pipelines and feature stores.
Optimize cloud resource utilization and cost efficiency for model training and inference workloads.
Real-Time & Streaming Use Cases (Preferred)
Support real-time or near real-time inference workflows using Kafka.
Design event-driven architectures for streaming model predictions and analytics.
Ensure low-latency, high-availability model serving in distributed systems.
Cross-Functional Collaboration
Partner with business stakeholders to translate requirements into AI-driven solutions.
Work closely with engineering teams to productionize models within enterprise applications.
Contribute to architecture decisions, technical documentation, and knowledge sharing.
Required Qualifications
Bachelor's or Master's degree in Data Science, Computer Science, Engineering, or related field (PhD preferred for advanced AI roles).
Strong hands-on experience in AI/ML model development and deployment.
Proficiency in TensorFlow and PyTorch for deep learning and generative AI model development.
Experience working with AWS cloud services for model training, deployment, and scaling.
Strong experience integrating ML workflows with Snowflake or enterprise data warehouse platforms.
Experience implementing MLOps practices, including CI/CD, model versioning, monitoring, and automation.
Knowledge of LLMOps principles for managing large language model lifecycle and governance.
Experience monitoring and optimizing model performance in production environments.
Familiarity with Kafka for real-time data streaming (preferred).
Preferred Qualifications
Experience deploying large language models and generative AI applications in production.
Knowledge of containerization and orchestration tools (e.g., Docker, Kubernetes).
Experience building feature stores and scalable inference APIs.
Understanding of AI ethics, bias mitigation, and responsible AI practices.
Key Competencies
Strong analytical and problem-solving abilities
Ability to translate business challenges into scalable AI solutions
Excellent communication and stakeholder management skills
Continuous learning mindset with strong research orientation
Sprachkenntnisse
- English
Hinweis für Nutzer
Dieses Stellenangebot stammt von einer Partnerplattform von TieTalent. Klicken Sie auf „Jetzt Bewerben“, um Ihre Bewerbung direkt auf deren Website einzureichen.