Research Scientist Intern, Vision Language Model (PhD)

Redmond, WA

Meta

Giving people the power to build community and bring the world closer together

View all jobs at Meta

Apply now Apply later

Reality Labs Research is looking for an intern to help us develop the next generation assistance systems that guide the users in contextual and adaptive future AR/VR systems. In particular, we are seeking candidates who have experience with either of the following: vision language model, LLM interpretability, multimodal LLM.

Work with researchers to help enable their work across the following research disciplines:
- Improving the performance of VLM in product-related scenarios
- Building white-box mechanisms to better evaluate the capabilities of VLMs

Our internships are twelve (12) to twenty-four (24) weeks long and we have various start dates throughout the year. Some projects may require a minimum of 16 consecutive weeks.Research Scientist Intern, Vision Language Model (PhD) Responsibilities
  • Develop, implement, and evaluate methods for improving the performance/interpretability of VLMs.
  • Make use of Meta’s large infrastructure to scale and speed up experimentation.
  • Write modular research code that can be reused in other contexts.
  • Collaborate with other researchers.
  • Work towards taking on big problems and deliver clear, compelling, and creative solutions to solve them at scale.
  • The work should result in publishable research to appear in a top-tier ML or CV conference (e.g., NeurIPS, ICLR, CVPR, ECCV).
Minimum Qualifications
  • Currently has or is in the process of pursuing a PhD in machine learning, computer vision, speech processing, applied statistics, computational neuroscience, or relevant technical field.
  • Excellent research skills involving defining problems, exploring solutions, and analyzing and presenting results.
  • Proficiency in python and machine learning libraries (pytorch, numpy, scikit-learn, scipy, pandas, matplotlib, etc.).
  • Deep understanding of vision-language models, supported by quality first-authored publications in related domains.
  • Interpersonal skills: cross-group collaboration and cross-culture collaboration.
  • Must obtain work authorization in the country of employment at the time of hire, and maintain ongoing work authorization during employment.
Preferred Qualifications
  • Proven track record of achieving significant results as demonstrated by grants, fellowships, patents, as well as first-authored publications at leading workshops or conferences such as NeurIPS, ICML, ICLR, CHI, UIST, IMWUT, CVPR, ICCV, ECCV, AAAI, ICRA, SIGGRAPH, ETRA, or similar.
  • Experience with VLM/LLM training/fine-tuning.
  • Experience on solving traditional CV problems, including but not limited to hand/body pose estimation, object detection, image classification/segmentation, image/video understanding, etc.
  • Experience working and communicating cross functionally in a team environment.
  • Intent to return to degree program after the completion of the internship/co-op.
  • Availability for minimum 16 consecutive week internship.
For those who live in or expect to work from California if hired for this position, please click here for additional information. LocationsAbout Meta Meta builds technologies that help people connect, find communities, and grow businesses. When Facebook launched in 2004, it changed the way people connect. Apps like Messenger, Instagram and WhatsApp further empowered billions around the world. Now, Meta is moving beyond 2D screens toward immersive experiences like augmented and virtual reality to help build the next evolution in social technology. People who choose to build their careers by building with us at Meta help shape a future that will take us beyond what digital connection makes possible today—beyond the constraints of screens, the limits of distance, and even the rules of physics.
$7,800/month to $11,293/month + benefits

Individual compensation is determined by skills, qualifications, experience, and location. Compensation details listed in this posting reflect the base hourly rate, monthly rate, or annual salary only, and do not include bonus, equity or sales incentives, if applicable. In addition to base compensation, Meta offers benefits. Learn more about benefits at Meta.

Equal Employment Opportunity and Affirmative Action Meta is proud to be an Equal Employment Opportunity and Affirmative Action employer. We do not discriminate based upon race, religion, color, national origin, sex (including pregnancy, childbirth, reproductive health decisions, or related medical conditions), sexual orientation, gender identity, gender expression, age, status as a protected veteran, status as an individual with a disability, genetic information, political views or activity, or other applicable legally protected characteristics. You may view our Equal Employment Opportunity notice here.
Meta is committed to providing reasonable support (called accommodations) in our recruiting processes for candidates with disabilities, long term conditions, mental health conditions or sincerely held religious beliefs, or who are neurodivergent or require pregnancy-related support. If you need support, please reach out to accommodations-ext@fb.com.
Apply now Apply later
Job stats:  0  0  0

Tags: Classification Computer Vision ICLR ICML LLMs Machine Learning Matplotlib NeurIPS NumPy Pandas PhD Physics Python PyTorch Research Scikit-learn SciPy Statistics VR

Perks/benefits: Career development Conferences Equity / stock options Health care Salary bonus Startup environment

Region: North America
Country: United States

More jobs like this