aijobs.net

Research Engineer - LLM/VLM Inference Optimization (Seed Infra)

San Jose, California, United States

USD 244K-450K Mid-level Full Time

Apply Save
Found 6h ago
Tasks
Perks/Benefits
Skills/Tech-stack

CUDA | Compiler optimization | Graph optimization | High concurrency | Low-precision computing | Parallel Computing | Performance Analysis | Precision computing | Speculative decoding | Streaming inference

Education

N/A

Roles

Engineer | Research Engineer

Regions

North America

Countries

United States

States

California, US

Cities

San Jose, California, US

Apply Save
Language: en | Views: 0 | Clicks: 0 | Saves: 0

Related jobs