SWE - Machine Learning Researcher, Foundation Models

Apple

Job Summary

We are a group of engineers and researchers responsible for building foundation models at Apple, including infrastructure, datasets, and models with fundamental general capabilities such as understanding and generation of text, images, speech, videos, and other modalities. We apply these models to Apple products. We are looking for researchers passionate about developing algorithms, techniques, and systems that push the frontier of deep learning and delight millions of users. You will work with a close-knit team to tackle challenging problems in foundation models and deep learning, identifying and developing novel applications.

Must Have

  • Build foundation models at Apple
  • Develop algorithms, techniques, and systems for deep learning
  • Focus on pretraining, LLM architecture, and scientific scaling of LLM
  • Demonstrated expertise in deep learning with publication record or product application
  • Proficient in Python and deep learning toolkits (JAX, PyTorch, or Tensorflow)
  • Ability to work in a collaborative environment
  • PhD, or equivalent practical experience, in Computer Science or related technical field

Good to Have

  • Experience in full-stack LLM optimization (mid-training, reinforcement learning, data research, kernel optimization)
  • Experience with pallas and triton
  • Code large language models
  • Reinforcement learning, on-policy distillation
  • Post-training, mid-training large language models
  • LLM context lengthening

Perks & Benefits

  • Comprehensive medical and dental coverage
  • Retirement benefits
  • Discounted products and free services
  • Reimbursement for certain educational expenses (tuition)
  • Opportunity to become an Apple shareholder through discretionary employee stock programs
  • Eligibility for discretionary restricted stock unit awards
  • Option to purchase Apple stock at a discount via Employee Stock Purchase Plan
  • Eligibility for discretionary bonuses or commission payments
  • Eligibility for relocation

Job Description

We are a group of engineers and researchers responsible for building foundation models. We build infrastructure, datasets, and models with fundamental general capabilities such as understanding and generation of text, images, speech, videos, and other modalities and apply these models to products. We are looking for researchers who are passionate about developing algorithms, techniques, and systems that push the frontier of deep learning and delight millions of users with products powered by foundation models.

We believe that the most interesting problems in deep learning research arise when we try to apply learning to real-world use cases, and this is also where the most important breakthroughs come from. You will work with a close-knit and fast growing team of world-class engineers and scientists to tackle some of the most challenging problems in foundation models and deep learning. Further, you will have opportunities to identify and develop novel applications of deep learning in products. You will see your ideas improve the experience of billions of users.

  • In this role, you will focus on pretraining, large language model (LLM) architecture, and scientific scaling of LLM. Experiences on full-stack LLM optimization such as mid-training, reinforcement learning, data research and kernel optimization (e.g. pallas and triton) will be a plus.
  • Demonstrated expertise in deep learning with publication record in relevant conferences (e.g., NeurIPS, ICML, ICLR, COLM, ACL, NAACL, EMNLP, ACL) or a track record in applying deep learning techniques to products
  • Proficient programming skills in Python and one of the deep learning toolkits such as JAX, PyTorch, or Tensorflow
  • Ability to work in a collaborative environment.
  • PhD, or equivalent practical experience, in Computer Science, or related technical field.
  • Code large language models.
  • Reinforcement learning, on-policy distillation.
  • Post-training, mid-training large language models.
  • LLM context lengthening.

At Apple, base pay is one part of our total compensation package and is determined within a range. This provides the opportunity to progress as you grow and develop within a role. The base pay range for this role is between $147,400 and $272,100, and your base pay will depend on your skills, qualifications, experience, and location.

Employees also have the opportunity to become a shareholder through participation in discretionary employee stock programs. Employees are eligible for discretionary restricted stock unit awards, and can purchase stock at a discount if voluntarily participating in the Employee Stock Purchase Plan. You’ll also receive benefits including: Comprehensive medical and dental coverage, retirement benefits, a range of discounted products and free services, and for formal education related to advancing your career, reimbursement for certain educational expenses — including tuition. Additionally, this role might be eligible for discretionary bonuses or commission payments as well as relocation. Learn more about Benefits.

Note: Benefit, compensation and employee stock programs are subject to eligibility requirements and other terms of the applicable plan or program.

Apple is an equal opportunity employer that is committed to inclusion and diversity. We seek to promote equal opportunity for all applicants without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability, Veteran status, or other legally protected characteristics. Learn more about your EEO rights as an applicant

.

Apple accepts applications to this posting on an ongoing basis.

9 Skills Required For This Role

Game Texts Html Pytorch Deep Learning Reinforcement Learning Python Algorithms Tensorflow Machine Learning

Similar Jobs