Technical Program Manager, Inference
San Francisco, CA
Full Time Mid-level / Intermediate USD 290K - 365K
Anthropic
Anthropic is an AI safety and research company that's working to build reliable, interpretable, and steerable AI systems.About Anthropic
Anthropic’s mission is to create reliable, interpretable, and steerable AI systems. We want AI to be safe and beneficial for our users and for society as a whole. Our team is a quickly growing group of committed researchers, engineers, policy experts, and business leaders working together to build beneficial AI systems.
About the role
As a Technical Program Manager for Inference, you'll drive critical initiatives that connect our cutting-edge AI research to production deployment at scale. You'll work closely with research teams to understand new model capabilities and requirements, then coordinate with infrastructure teams to deploy these models reliably at scale. This role is essential in orchestrating complex model deployments, understanding how inference decisions impact capacity, and balancing the trade-offs between latency, throughput, and cost across our infrastructure.
You'll collaborate with teams working on diverse platforms including TPUs, GPUs, and Trainium, helping to optimize our inference stack for both current production models and future model architectures. This position offers unique challenges in scaling AI systems while maintaining the reliability and performance our customers expect.
Responsibilities:
- Lead end-to-end program management for model bring-up and deployment, coordinating between research, infrastructure, and platform teams
- Manage complex technical programs involving model optimization, capacity planning, and multi-cloud deployment strategies
- Drive cross-functional initiatives for major launches, collaborating with partner-facing teams to ensure external dependencies are met
- Track and communicate critical metrics including capacity utilization, latency targets, and reliability SLOs across different serving platforms
- Navigate technical trade-offs between performance, cost, and reliability, helping teams make informed decisions about inference optimizations
- Coordinate with research teams to understand upcoming model architectures and prepare infrastructure accordingly
- Lead technical discussions around serving strategies for different model families and deployment platforms
- Partner with engineering teams to improve inference reliability and help meet our SLA targets
- Create and maintain comprehensive documentation of deployment processes, capacity models, and technical decisions
You may be a good fit if you:
- Have several years of experience in technical program management, with a track record of successfully delivering complex infrastructure or ML systems
- Possess deep technical knowledge that allows you to engage meaningfully with ML researchers and infrastructure engineers
- Have experience with large-scale distributed systems and understand the challenges of serving ML models in production
- Are comfortable working with multiple cloud platforms and hardware accelerators
- Excel at translating between technical teams and stakeholders, making complex trade-offs understandable
- Have experience managing programs with significant external dependencies and multiple stakeholder groups
- Can thrive in ambiguous situations, bringing structure to complex technical challenges
- Have strong analytical skills and can work with performance metrics, capacity models, and cost optimization
- Are passionate about AI technology and interested in the challenges of deploying frontier models at scale
- Have excellent written and verbal communication skills, with the ability to influence without authority
The expected salary range for this position is:
Annual Salary:$290,000—$365,000 USDLogistics
Education requirements: We require at least a Bachelor's degree in a related field or equivalent experience.
Location-based hybrid policy: Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.
Visa sponsorship: We do sponsor visas! However, we aren't able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.
We encourage you to apply even if you do not believe you meet every single qualification. Not all strong candidates will meet every single qualification as listed. Research shows that people who identify as being from underrepresented groups are more prone to experiencing imposter syndrome and doubting the strength of their candidacy, so we urge you not to exclude yourself prematurely and to submit an application if you're interested in this work. We think AI systems like the ones we're building have enormous social and ethical implications. We think this makes representation even more important, and we strive to include a range of diverse perspectives on our team.
How we're different
We believe that the highest-impact AI research will be big science. At Anthropic we work as a single cohesive team on just a few large-scale research efforts. And we value impact — advancing our long-term goals of steerable, trustworthy AI — rather than work on smaller and more specific puzzles. We view AI research as an empirical science, which has as much in common with physics and biology as with traditional efforts in computer science. We're an extremely collaborative group, and we host frequent research discussions to ensure that we are pursuing the highest-impact work at any given time. As such, we greatly value communication skills.
The easiest way to understand our research directions is to read our recent research. This research continues many of the directions our team worked on prior to Anthropic, including: GPT-3, Circuit-Based Interpretability, Multimodal Neurons, Scaling Laws, AI & Compute, Concrete Problems in AI Safety, and Learning from Human Preferences.
Come work with us!
Anthropic is a public benefit corporation headquartered in San Francisco. We offer competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lovely office space in which to collaborate with colleagues.
Tags: Anthropic Architecture Biology Computer Science Distributed Systems Engineering Excel GPT GPT-3 Machine Learning ML models Physics Research
Perks/benefits: Career development Competitive pay Equity / stock options Flex hours Flex vacation Parental leave
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.