1x Product + 1x ML Staff Engineer : Santa Clara, CA (HQ)
5/5/26About 1 min
VLM Run (https://vlm.run) - 1x Product + 1x ML Staff Engineer - Santa Clara, CA (HQ)
We're building the inference and orchestration layer for production Vision-Language Models. We care deeply about fast and ergonomic visual inference, reliable structured outputs, and the observability to iterate on them.
A few things we've shipped recently you can poke at:
1. Orion: our visual agent that reasons and acts over images, video, and documents. Chat at https://chat.vlm.run.
2. mm-ctx: a Unix-style multimodal CLI (find, cat, grep, wc) that gives coding agents real context over images, video, and PDFs. Rust core, Python devex.
3. vlmbench: single-file CLI for benchmarking VLM inference (TTFT, TPOT, throughput) across vLLM, Ollama, and SGLang.
Apply: https://app.dover.com/jobs/vlm-run
Email hiring "at" vlm.run with your GitHub + a couple recent projects.
[2] https://pypi.org/project/mm-ctx - https://www.vlm.run/open-source/mm
[3] https://github.com/vlm-run/vlmbench - https://www.vlm.run/open-source/vlmbench
Need focus music?Muz11 - 100+ curated playlists for coding
Compare AI models?Sneos - Chat with multiple AIs side by side
Love coffee?Want to switch careers? Check Career.Coffee!
Need an AI agent?NestClaw - Private AI agent on a dedicated server
Find tech jobs?Career.Computer - Tech jobs for tomorrow's innovators
Apply for this Role
Your application will be forwarded to the hiring contact.