aijobs.net

Research Engineer - LLM/VLM Inference Optimization (Seed Infra)

San Jose, California, United States

USD 244K-450K Mid-level Full Time

Apply Save
Found 24d ago
Tasks
Perks/Benefits
Skills/Tech-stack

CUDA | Compiler optimization | Graph optimization | High concurrency | Low-precision computing | Parallel Computing | Performance Analysis | Precision computing | Speculative decoding | Streaming inference

Education

N/A

Roles

Engineer | Research Engineer

Regions

North America

Countries

United States

States

California, US

Cities

San Jose, California, US

Apply Save
Language: en Views: 1 Clicks: 0 Saves: 0

Related jobs