AI Research Scientist - Foundation Models & Multimodal Intelligence at Apple
Sunnyvale, California, USA -
Full Time


Start Date

Immediate

Expiry Date

12 Nov, 25

Salary

272100.0

Posted On

12 Aug, 25

Experience

3 year(s) or above

Remote Job

Yes

Telecommute

Yes

Sponsor Visa

No

Skills

Python, Machine Learning, Fine Tuning, Computer Science, Personalization, Deep Learning, Collaborative Environment, Distributed Systems, Data Curation, Production Deployment, Learning Techniques

Industry

Information Technology/IT

Description

Imagine building the next generation of AI-powered experiences at Apple. We are advancing the state of the art in foundation models, applying them across language, vision, and multimodal understanding to power features used by millions of people worldwide. As part of the Multimodal Intelligence Team (MINT), with a track record of delivering innovations from the Apple Foundation Model to real-world applications like Visual Intelligence, you will tackle the practical challenges of scaling, optimizing, for building large models as well as integrating such models and agents into Apple products. You’ll collaborate with world-class engineers and scientists to push the boundaries of foundation models and agentic system while delivering real-world impact.

DESCRIPTION

You will work on advancing the capabilities of foundation models and steering them toward real-world applications in Apple products. This includes researching and developing methods to improve alignment, reasoning, and adaptation of large models to practical use cases, while ensuring they meet Apple’s standards for efficiency, scalability, and privacy. A key part of your role will be to stay current with emerging research and identify techniques suitable for real-world deployment, bridging the gap between state-of-the-art advancements and production-ready solutions. You will design and optimize large-scale data pipelines to enable robust training and fine-grained evaluation of foundation models, working with massive multimodal datasets to push the limits of performance. You will explore new ways to enhance reasoning and multimodal understanding, adapting models for Apple’s unique ecosystem, from cloud-scale infrastructure to on-device intelligence. Collaboration will be central to your work-you will partner with cross-functional teams of engineers and researchers bring these models to life, ensuring seamless integration into Apple products and creating intelligent natural user experiences.

MINIMUM QUALIFICATIONS

  • BS and a minimum of 3 years relevant industry experience.
  • Proficient programming skills in Python and experience with at least one modern deep learning framework (PyTorch, JAX, or TensorFlow).
  • Experience working with large-scale training pipelines and distributed systems.

PREFERRED QUALIFICATIONS

  • PhD, or equivalent practical experience, in Computer Science, Machine Learning, or a related technical field.
  • Demonstrated expertise in deep learning, with either: A publication record in relevant conferences (e.g., NeurIPS, ICML, ICLR, CVPR, ICCV, ECCV,COLM, etc), or a strong track record of applying deep learning techniques to real-world products.
  • Experience with foundation models (language, vision-language, or multimodal).
  • Familiarity with large-scale data pipelines, including data curation, preprocessing, and efficient storage.
  • Experience fine-tuning or optimizing large models for production deployment.
  • Knowledge of retrieval-augmented generation (RAG), personalization, or grounding techniques.
  • Familiarity with privacy-preserving or on-device machine learning.
  • Ability to work effectively in a cross-functional, collaborative environment.

How To Apply:

Incase you would like to apply to this job directly from the source, please click here

Responsibilities

Please refer the Job description for details

Loading...