À propos
Senior Machine Learning Research Engineer (Multi-Sensor Fusion Perception & Foundation Models)
to join our Berkeley office. Zendar develops one of the best 360-degree radar-based vehicular perception systems for automotive. We're now expanding our capabilities to deliver full-scene perception outputs using
early fusion of camera and radar , and scaling these technologies across both the automotive and robotics industries. We are not bogged down by legacy systems, and by joining us you'll have the opportunity to
define and own a next-generation perception stack
that enables reliable autonomy at scale.
About Zendar:
Zendar is building perception for physical AI-giving engineers a strong foundation for creating world-class robotics applications. At Zendar, you'll work on perception foundation models that enable robots to understand and interact with their environments across a wide range of industries.
Zendar pioneered RF perception that delivers a vision-like, semantically segmented understanding of the environment-running on embedded automotive systems using only radar data. This RF perception forms the backbone of Zendar's next-generation foundation models, which are built around early fusion of RF and vision data.
This architecture inverts the traditional perception stack. Instead of treating RF signals as secondary, Zendar's models combine vision's high angular resolution with RF's strong temporal and spatial understanding at the earliest stages of perception. The result is a system that sees farther, remains robust to occlusion and adverse weather, and operates far more efficiently than vision-only or lidar-based approaches.
See a demo of Zendar's foundational RF perception
At Zendar, you'll work at the cutting edge of autonomous mobility and robotics-advancing foundation models that will power the next generation of physical AI systems. You'll work with large-scale, real-world, multi-modal datasets composed of synchronized and calibrated radar, camera, and lidar data collected across multiple continents.
Our team brings together deep expertise across hardware, signal processing, machine learning, and software engineering, with decades of experience in sensing and perception. We are a global team with offices in Berkeley, Lindau (Germany), and Paris (France). Zendar is well-funded by leading Tier-1 venture capital firms and has established strong industry partnerships.
Although AI is central to what we build, our hiring process is intentionally human: every résumé is reviewed by a real person.
Your Role:
Zendar's Semantic Spectrum perception technology extracts a rich scene understanding from radar sensing. Our next goal is to build a foundation-model-driven perception stack that fuses streaming camera and radar to produce full perception outputs that are robust enough for real-world autonomy: occupancy/free-space (e.g., occupancy grid), object detection and tracking, lane line and road structure estimation, and the interfaces required to make these outputs actionable for downstream systems.We are seeking an experienced Senior ML Engineer to design, implement, and drive the architecture of these models end-to-end, including training from scratch on large-scale datasets (not just fine-tuning), defining evaluation and long-tail validation, and partnering with platform and product teams to ensure successful deployment in real-time systems.This is an ideal position for an engineer who enjoys owning hard technical problems, making rigorous tradeoffs, and building systems that work reliably in the messy long tail of the real world.In this role you will have close communication and collaboration with platform, embedded, and robotics teams. You will work with our real-world dataset of tens of thousands of kilometers collected in multiple continents and geographies, and you will have opportunities to validate results on real vehicles.
What You'll Do:
Own architecture and technical strategy for multi-sensor perception models, including explicit tradeoffs (why approach A vs B), risks, validation plans, and timelines.Build foundation-scale / transformer-based perception models trained from scratch on large-scale multi-modal driving datasets (not limited to fine-tuning).Develop fusion architectures for streaming multi-sensor inputs (camera/radar/lidar), with early fusion and temporal fusion; align training objectives to real-world reliability targets.Deliver production-ready models for:Occupancy / free-space / dynamic occupancy (full-scene understanding)3D Object detection and trackingLane line / road structure estimationDrive long-tail reliability (e.g., toward "four nines" behavior in defined conditions)Partner with platform/embedded teams to ensure models meet real-time constraints (latency, memory, throughput) and integrate cleanly via stable interfaces for downstream consumers.
What We Look For:
Deep expertise in deep learning for perception, especially
transformer-based
architectures, temporal modeling, and multi-modal learning. Proficiency with Python and a major deep learning framework (e.g., PyTorch, TensorFlow) 5+ years (or having a PhD) experience designing and implementing ML systems, with demonstrated ownership of research/production outcomes. Demonstrated experience training
large models from scratch
(not only fine-tuning) Strong experience with
multi-sensor fusion
(camera/radar/lidar) and real-world sensor Strong understanding of the end-to-end perception stack and downstream needs (interfaces, uncertainty, temporal stability, failure modes). Ability to lead architectural discussions: articulate tradeoffs, quantify risks/benefits, and set realistic milestones and timelines. Bonus Points: PhD
in a relevant field (Machine Learning, Computer Vision, Robotics) preferred. Experience with foundation models for autonomy and robotics, including multi-modal pretraining, self-supervised learning, and scaling laws / model scaling strategies. Experience with transfusion-style or related fusion paradigms (transformer-based fusion across modalities and time), including building from first principles. Experience with BEV-centric perception, 3D detection, occupancy networks, tracking, and streaming inference. What We Offer:
Opportunity to make an impact at a young, venture-backed company in an emerging market Competitive salary ranging from $140,000 to $190,000 annually depending on experience Performance based Bonus Benefits including medical, dental, and vision insurance, flexible PTO, and equity Hybrid work model: in office 3 days per week from Tuesday to Thursday, the rest... work from wherever! Daily catered lunch and a stocked fridge (when working out of the Berkeley, CA office)
Zendar is committed to creating a diverse environment where talented people come to do their best work. We are proud to be an equal opportunity employer. All qualified applicants will receive consideration for employment without regard to race, color, religion, gender, gender identity or expression, sexual orientation, national origin, genetics, disability, age, or veteran status.
Zendar participates in E-Verify
Compétences linguistiques
- English
Avis aux utilisateurs
Cette offre provient d’une plateforme partenaire de TieTalent. Cliquez sur « Postuler maintenant » pour soumettre votre candidature directement sur leur site.