Machine Learning Engineer — Distillation
Quick Summary
About the Role We’re looking for a Machine Learning Engineer focused on model distillation to help us build smaller, faster, and more efficient models without sacrificing quality.
Experience distilling LLMs or large sequence models Experience with inference optimization (quantization, pruning, kernels, etc.) Familiarity with evaluation for language models Open-source contributions or research publications Experience in…
About the Role
~1 min readWe’re looking for a Machine Learning Engineer focused on model distillation to help us build smaller, faster, and more efficient models without sacrificing quality. You’ll work at the intersection of research and production—taking cutting-edge techniques and turning them into systems that scale.
This is a hands-on role with real ownership: you’ll design distillation pipelines, run large-scale experiments, and ship models used in production.
Responsibilities
~1 min read- →
Design and implement knowledge distillation pipelines (teacher–student, self-distillation, multi-teacher, etc.)
- →
Distill large foundation models into smaller, faster, and cheaper models for inference
- →
Run and analyze large-scale training experiments to evaluate quality, latency, and cost tradeoffs
- →
Collaborate with research to translate new distillation ideas into production-ready code
- →
Optimize training and inference performance (memory, throughput, latency)
- →
Contribute to internal tooling, evaluation frameworks, and experiment tracking
- →
(Optional) Contribute back to open-source models, tooling, or research
Strong background in machine learning or deep learning
Hands-on experience with model distillation (LLMs or other neural networks)
Solid understanding of training dynamics, loss functions, and optimization
Experience with PyTorch (or JAX) and modern ML tooling
Comfort running experiments on multi-GPU or distributed setups
Ability to reason about model quality vs. performance tradeoffs
Pragmatic mindset: you care about shipping, not just papers
Nice to Have
~1 min readExperience distilling LLMs or large sequence models
Experience with inference optimization (quantization, pruning, kernels, etc.)
Familiarity with evaluation for language models
Open-source contributions or research publications
Experience in early-stage or fast-moving startups
What We Offer
~1 min readLocation & Eligibility
Listing Details
- Posted
- January 22, 2026
- First seen
- May 6, 2026
- Last seen
- May 10, 2026
Posting Health
- Days active
- 0
- Repost count
- 0
- Trust Level
- 23%
- Scored at
- May 6, 2026
Signal breakdown
Please let featherlessai know you found this job on Jobera.
4 other jobs at featherlessai
View all →Explore open roles at featherlessai.
Similar Machine Learning Engineer jobs
View all →Stay ahead of the market
Get the latest job openings, salary trends, and hiring insights delivered to your inbox every week.
No spam. Unsubscribe at any time.