Member of Technical Staff - Inference
Tasks
- Build autoscaling routing load balancing
- Build observability tools
- Build open superintelligence infrastructure
- Design scheduling algorithms
- Develop LLM serving platform
- Develop performance suites
- Document architectures and collaborate
- Embed inference into RL stack
- Establish CI/CD pipelines
- Implement resilience and failover
- Integrate and develop inference frameworks
- Optimize Inference Systems
- Optimize model distribution
- Profile and tune parallelism
Perks/Benefits
- Competitive compensation
- Conference attendance
- Equity incentives
- Flexible work
- Professional development
- Relocation support
- Remote option
- Team off-sites
Skills/Tech-stack
AWS | Ansible | Benchmarks | C++ | CUDA | Disaggregated serving | GCP | GRPC | Grafana | Heterogeneous scheduling | Infiniband | KVCache | Kafka | Kubernetes | Linux | NCCL | Nvidia Dynamo | OpenTelemetry | Performance Profiling | Prometheus | Protobuf | PubSub | PyTorch | Python | Quantization | Redis | Rust | Speculative decoding | TensorRT | Terraform
Education
Related jobs
-
Member of Technical Staff (AI Supportability) INR 2500K-5000KAnthropic API | Data Analysis | Distributed Systems | Go | Machine LearningCollaborative team environment | Learning and development opportunitiesSenior-level Full TimeBangalore, India R4d ago
-
Member of Technical Staff (AI Supportability) INR 2500K-5000KAI Model Evaluation | AI model | AI pattern research | APIs | Data-Driven ExperimentationCollaboration with technical teams | Opportunity to impact product strategy | Professional growth with cutting-edge technologies | Work in innovative AI environmentSenior-level Full TimeBangalore, India R4d ago