AI Summary
Lead BEV model development, design advanced multi-modal architectures, and drive architectural innovation for autonomous driving stack.
Key Highlights
Lead BEV model development
Design advanced multi-modal architectures
Drive architectural innovation
Develop foundational perception models
Mentor and guide ML engineers
Technical Skills Required
Benefits & Perks
Competitive compensation package
Bonus component
Stock options
Medical, dental, and vision insurance
RRSP plan with 4% employer match
Public Transit Subsidy
Flexibility in schedule
Generous paid vacation
Life Insurance
Job Description
About The Company
At Torc, we have always believed that autonomous vehicle technology will transform how we travel, move freight, and do business.
A leader in autonomous driving since 2007, Torc has spent over a decade commercializing our solutions with experienced partners. Now a part of the Daimler family, we are focused solely on developing software for automated trucks to transform how the world moves freight.
Join us and catapult your career with the company that helped pioneer autonomous technology, and the first AV software company with the vision to partner directly with a truck manufacturer.
Meet the Team:
As a Staff Machine Learning Engineer specializing in BEV (Bird’s-Eye View) and Multi-Modal Perception, you will lead the development of next-generation models that unify information across cameras, LiDAR and radar to deliver a rich spatial understanding of the driving environment.
You will drive architectural innovation, large-scale model training, and data-driven improvements that directly advance the perception capabilities at the heart of Torc’s autonomous driving stack.
This is a technical leadership role focused on model innovation and maturity, not downstream feature integration.
What You’ll Do
- Lead BEV model development: Define and execute the technical roadmap for BEV-based perception models across multiple tasks (e.g., detection, segmentation, road topology, and scene understanding).
- Design advanced multi-modal architectures that fuse heterogeneous sensor data (camera, LiDAR, radar, HD maps) into unified spatial representations.
- Develop foundational perception models leveraging BEV transformers, voxel-based encoders, or implicit scene representations.
- Own large-scale training workflows — from data sampling strategies and augmentation pipelines to distributed training and hyperparameter optimization.
- Advance model robustness and generalization, addressing long-tail conditions such as low visibility, occlusions, and rare scene configurations.
- Establish evaluation frameworks for geometric accuracy, temporal stability, and cross-domain transfer performance.
- Collaborate cross-functionally with sensor calibration, mapping, and fusion teams to ensure cohesive perception model interfaces.
- Mentor and guide ML engineers, cultivating best practices in experimentation, code quality, and model validation.
- Stay at the forefront of ML research, exploring self-supervised learning, large-scale pretraining, or foundation models for 3D perception.
- 10+ years of experience in deep learning for perception, 3D vision, and/or autonomous systems
- M.S. or Ph.D. in Computer Science, Electrical Engineering, Robotics, or related field (or equivalent practical experience).
- Proven expertise in BEV modeling, 3D scene understanding, and multi-view fusion.
- Strong background in multi-modal sensor fusion, particularly integrating camera and LiDAR data.
- Proficiency in Python and deep learning frameworks such as PyTorch or TensorFlow.
- Experience with large-scale data pipelines, distributed training, and experiment management systems.
- Demonstrated leadership in driving ML model innovation and mentoring technical teams.
- Experience with autonomous driving or robotics perception in production environments.
- Experience with MLOps and infrastructure tools (Ray)
- Hands-on expertise in BEV-based ML architectures, LiDAR-vision fusion, or spatial-temporal modeling.
- Familiarity with 3D labeling, calibration, and sensor simulation pipelines.
- Track record of publications or open-source contributions in top-tier venues (CVPR, ICCV, NeurIPS, ICRA, CoRL).
- Understanding of performance tradeoffs and deployment constraints (latency, memory, accuracy).
Work Location: For this position, we are open to hiring in either the Torc Montreal, Quebec (Canada) or Ann Arbor, MI (U.S.) office work locations in a hybrid capacity. We are also open to hiring Remote in the United States or Canada.
Perks of Being a Full-time Torc’r (Canada)
Torc cares about our team members and we strive to provide benefits and resources to support their health, work/life balance, and future. Our culture is collaborative, energetic, and team focused. Torc offers:
- A competitive compensation package that includes a bonus component and stock options
- Medical, dental, and vision for full-time employees
- RRSP plan with a 4% employer match
- Public Transit Subsidy (Montreal area only)
- Flexibility in schedule and generous paid vacation
- Company-wide holiday office closures
- Life Insurance
Our compensation reflects the cost of labor across several geographic markets. Pay is based on a number of factors and may vary depending on job-related knowledge, skills, and experience. Torc's total compensation package will also include our corporate bonus and stock option plan. Dependent on the position offered, sign-on payments, relocation, and other forms of compensation may be provided as part of a total compensation package, in addition to a full range of medical, financial, and/or other benefits.
CAD Compensation Range: $209,300-313,800 CAD
Job ID: 102405