featherlessai
New

Machine Learning Engineer — Distillation

(world)Remotefull-timemid
Machine Learning EngineerData
0 views0 saves0 applied

Quick Summary

Overview

About the Role We’re looking for a Machine Learning Engineer focused on model distillation to help us build smaller, faster, and more efficient models without sacrificing quality.

Requirements Summary

Experience distilling LLMs or large sequence models Experience with inference optimization (quantization, pruning, kernels, etc.) Familiarity with evaluation for language models Open-source contributions or research publications Experience in…

Technical Tools
pytorchdeep-learningmachine-learning

About the Role

~1 min read

We’re looking for a Machine Learning Engineer focused on model distillation to help us build smaller, faster, and more efficient models without sacrificing quality. You’ll work at the intersection of research and production—taking cutting-edge techniques and turning them into systems that scale.

This is a hands-on role with real ownership: you’ll design distillation pipelines, run large-scale experiments, and ship models used in production.

Responsibilities

~1 min read
  • Design and implement knowledge distillation pipelines (teacher–student, self-distillation, multi-teacher, etc.)

  • Distill large foundation models into smaller, faster, and cheaper models for inference

  • Run and analyze large-scale training experiments to evaluate quality, latency, and cost tradeoffs

  • Collaborate with research to translate new distillation ideas into production-ready code

  • Optimize training and inference performance (memory, throughput, latency)

  • Contribute to internal tooling, evaluation frameworks, and experiment tracking

  • (Optional) Contribute back to open-source models, tooling, or research

  • Strong background in machine learning or deep learning

  • Hands-on experience with model distillation (LLMs or other neural networks)

  • Solid understanding of training dynamics, loss functions, and optimization

  • Experience with PyTorch (or JAX) and modern ML tooling

  • Comfort running experiments on multi-GPU or distributed setups

  • Ability to reason about model quality vs. performance tradeoffs

  • Pragmatic mindset: you care about shipping, not just papers

Nice to Have

~1 min read
  • Experience distilling LLMs or large sequence models

  • Experience with inference optimization (quantization, pruning, kernels, etc.)

  • Familiarity with evaluation for language models

  • Open-source contributions or research publications

  • Experience in early-stage or fast-moving startups

What We Offer

~1 min read
Work on core model quality and cost efficiency—not side projects
High ownership and direct impact on product and roadmap
Small, senior team with strong research + engineering culture
Competitive compensation + meaningful equity
Remote-friendly, async-first environment

Location & Eligibility

Where is the job
Worldwide
Fully remote, anywhere in the world
Who can apply
Same as job location

Listing Details

Posted
January 22, 2026
First seen
May 6, 2026
Last seen
May 10, 2026

Posting Health

Days active
0
Repost count
0
Trust Level
23%
Scored at
May 6, 2026

Signal breakdown

freshnesssource trustcontent trustemployer trust
Newsletter

Stay ahead of the market

Get the latest job openings, salary trends, and hiring insights delivered to your inbox every week.

A
B
C
D
Join 12,000+ marketers

No spam. Unsubscribe at any time.

featherlessaiMachine Learning Engineer — Distillation
Machine Learning Engineer — Distillation | featherlessai | Remote (world)