Member of Technical Staff - Agentic Systems
Palo Alto
Context Labs
TL;DR
Context is turning large language models into autonomous co‑workers. Our productivity suite lets AI plan, reason, and execute across 300+ business tools. We've raised $11M (Seed, Lux Capital) at a $70M valuation and ship GA this June. We're hiring engineers who live and breathe LLMs—building the SDKs, runtimes, and eval harnesses that make LLM agents reliable at scale.
If you sketch agent flows in Excalidraw, obsess over token‑level traces, etc., let's talk.
Why build agents at Context?
Impact: Frontier playground
50M‑token memory → agents with persistent cognition well beyond frontier models.
Full‑stack autonomy: Agents open docs, write slides, hit APIs, and re‑plan—no brittle RPA hacks.
Tight feedback loops: Beta fleet drives >5K agent runs/day; your PRs hit prod within hours.
Elite peers: Microsoft Research, Stanford, Tesla peers—zero bureaucracy.
Hardware ally: Qualcomm Snapdragon X Elite partnership: 10× lower latency, on‑device privacy. Gateway into enterprise & government.
What you'll build
Agent SDK (TypeScript) — a composable toolkit (task graph, tools, memory adapters) powering both server‑side and edge runtimes.
Swarm Runtime — orchestrator in Node.js/Workers that schedules, supervises, and retries agent sub‑tasks at millisecond granularity.
Graph Traversal APIs — ergonomic wrappers around our 50M‑token knowledge graph with streaming, partial updates, and vector fallbacks.
Adaptive Prompt Compiler — real‑time TS library that compiles structured inputs → model‑specific prompts and parses responses.
Evaluation Harness — Jest/Playwright‑style test runner that auto‑red‑teams agents for hallucinations, tool‑use failure, and latency regressions.
Tech you'll touch
TypeScript • Node.js (18+) • Bun • Cloudflare Workers / V8 isolate
LangChainJS, LlamaIndex TS, custom lightweight frameworks
OpenAI, Anthropic, Llama CPP bindings
Rust (via NAPI) for hot paths • WebGPU/WebAssembly for client‑side reasoning
Postgres • Redis • Apache Arrow • bespoke graph store
Kubernetes • GCP • Edge functions
You might be a fit if
You've shipped production TS/Node services at scale and love type‑level safety.
You've built or contributed to agent / retrieval frameworks (LangChainJS, AutoGen, custom) and can dissect their failure modes.
You treat prompts as code: unit‑test them, version them, and lint them.
Comfortable jumping into Rust/Wasm when the profiler says so.
You think debuggers, trace viewers, and live‑reload matter as much as model choice.
Bonus points
Authored OSS in the LLM tooling ecosystem (LangChainJS plugin, OpenAI helper lib, etc.).
Experience with CRDT/OT for collaborative agent state or live‑share UIs.
Prior work on distributed schedulers or graph databases.
Compensation & Perks
Top‑of‑band salary + founder‑level equity
Medical, dental, vision • 401(k) match
Custom rig & conference budget
Bi‑annual team off‑sites (Tahoe powder, NYC AI Week, etc.)
Ready to push LLMs beyond chat—with TypeScript?
Email team@context.inc with a short intro and links (repos, demos, papers). We reply within 48 hours.
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Tags: Anthropic APIs Arrow GCP Kubernetes LLaMA LLMs Node.js OpenAI Playwright PostgreSQL Privacy Research Robotics RPA Rust Streaming TypeScript
Perks/benefits: 401(k) matching Equity / stock options Health care Salary bonus
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.