Skip to main content
V

Backend Engineer - Inference Optimization

Vercept
Full Timemid
Seattle, Washington, US$150k – $250kPosted March 3, 2026

Salary Context

This role offers $150k–$250k. The median for Mid-level backend roles is $100k–$160k (based on 44 listings). 54% above median.

Job Description

About Us

We're a high-energy, impact-driven team, with a long track record of academic excellence. Our team includes researchers whose work has shaped the field-earning best paper awards at top AI conferences and even ranking among the most cited scientists in history of science. We've built fundamental, transformative research that has redefined the community, and now, we're here to change the world-one breakthrough at a time.

What We're Looking For & Why Join Us

We're looking for a Backend Engineer - Inference Optimization who thrives on solving some of the hardest systems problems in AI. You'll focus on pushing the limits of foundation model inference performance, working at the intersection of cutting-edge ML and high-performance systems engineering. This is your opportunity to set new benchmarks for latency, throughput, and efficiency at scale.

What is this role?

As a Backend Engineer, you'll own the design and optimization of inference pipelines for large-scale models. You'll work closely with researchers and infrastructure engineers to identify bottlenecks, implement advanced techniques like quantization and KV caching, and deploy high-performance serving systems in production. Your work will directly determine how fast and cost-effectively users can access next-generation AI.

What do we expect?

Must have:

  • Deep experience in optimizing model inference pipelines, model quantization and KV caching.
  • Proficiency in backend systems and high-performance programming (Python, C++, or Rust)
  • Familiarity with distributed serving, GPU acceleration, and large-scale systems
  • Ability to debug complex performance issues across model, runtime, and hardware layers
  • Comfort working in fast-moving environments with ambitious technical goals

Nice to have

  • Hands-on experience with vLLM or similar inference frameworks
  • Background in GPU kernel optimization (CUDA, Triton, ROCm)
  • Experience scaling inference across multi-node or heterogeneous clusters
  • Prior work in model compilation (e.g., TensorRT, TVM, ONNX Runtime)
  • Hands-on experience with model quantization

Compensation & Benefits

$150K - $250K + Equity

We offer health benefits, a 401(k) plan, and meaningful equity-because we believe top talent should be supported, secure, and fully invested in the future we're building together.

Location: Our company is in-office at our Seattle HQ.

Want AI-powered job matching?

Upload your resume and get every job scored, your resume tailored, and hiring manager emails found - automatically.

Get Started Free