Quick Summary
Distillation for generative models (diffusion, autoregressive, flow-m
- Define and drive the technical strategy for model distillation and compression across Waabi's AI stack — spanning perception, world models, and planning — with an eye toward both onboard deployment and simulation use-cases.
- Design, implement, and scale state-of-the-art distillation and efficiency pipelines, which may include:
-
Distillation for generative models (diffusion, autoregressive, flow-matching, video models)
-
Quantization-aware training (QAT) and post-training quantization (PTQ)
-
Knowledge distillation (feature-level, response-based, and relation-based)
-
Structured and unstructured pruning and sparsification
-
Low-rank factorization and efficient architecture design
-
Speculative decoding and other inference-time efficiency techniques
- Collaborate closely with ML Platform, Infrastructure, Onboard, Autonomy, and Simulation teams to integrate compressed models into production pipelines and meet latency, memory, and throughput targets across deployment contexts.
- Define rigorous benchmarks and evaluation frameworks to characterize efficiency vs. quality trade-offs across models and hardware targets.
- Mentor and guide researchers and engineers working in the distillation and model efficiency space, setting a high technical bar and fostering a culture of rigorous experimentation.
- Champion best practices for model compression across the organization; disseminate knowledge through internal design reviews, documentation, and technical talks.
- Stay at the cutting edge of model efficiency research; contribute to the broader scientific community through publications and open-source contributions.
Requirements
~1 min readNice to Have
~1 min read- Experience with hardware-aware optimization (TensorRT, ONNX, custom CUDA kernels, hardware-specific quantization).
- Publications at top-tier ML/CV venues (NeurIPS, ICML, CVPR, ICLR, ECCV) in model compression, efficient deep learning, or related areas.
- Experience distilling large generative models (diffusion models, LLMs, VLMs, or video models).
- Background in autonomous vehicles or robotics.
Location & Eligibility
Listing Details
- Posted
- April 30, 2026
- First seen
- April 30, 2026
- Last seen
- May 4, 2026
Posting Health
- Days active
- 3
- Repost count
- 0
- Trust Level
- 70%
- Scored at
- May 4, 2026
Signal breakdown

Waabi is an AI company developing generative AI-powered autonomous driving technology for long-haul trucking, aiming for commercial deployment in 2025.
View company profilePlease let Waabi know you found this job on Jobera.
3 other jobs at Waabi
View all →Explore open roles at Waabi.
Browse Similar Jobs
Stay ahead of the market
Get the latest job openings, salary trends, and hiring insights delivered to your inbox every week.
No spam. Unsubscribe at any time.