AI Software Solutions Engineer (AI Frameworks, Workloads)

4 Weeks ago • 10-15 Years • Artificial Intelligence • Research & Development

About the job

Job Description

As an AI Software Solutions Engineer, you'll be responsible for designing, developing, and optimizing AI frameworks for inference. Your role will involve collaborating with cross-geo teams to enhance the inference stack, ensuring competitive performance on deep learning inference models with a specific focus on the PyTorch framework. Key responsibilities include designing and developing software techniques for AI frameworks, both hardware-agnostic and hardware-aware, contributing to the enhancement and expansion of Inference and Training capabilities in the software stack, profiling deep learning inference workloads to identify optimization opportunities, and implementing optimizations for deployment of Large Language Models (LLMs). You'll need expertise in advanced C++, Python, and parallel programming, along with a deep understanding of machine learning systems optimization and deployment techniques, including quantization. Experience with transformers and inference-specific optimizations is crucial, as is the ability to debug complex issues in multi-layered software systems. Strong understanding of software integration across open-source frameworks and internal framework layers, computer architecture, and effective communication skills are essential. Experience working in a cross-geo setup is preferred.
Must have:
  • BTech, MS or PhD in CS or related fields
  • 2-3 years of experience working on Inference frameworks
  • Architecture/Design contributions to Inference systems
  • Detailed understanding of machine learning systems optimization
  • Experience with optimization techniques for deployment of LLMs
  • Deep implementation knowledge of transformers
  • Programming skills in Advanced C++, Python, and parallel programming
  • Understanding of SW integration across open source frameworks
  • Strong understanding of computer architecture
  • Effective communication skills
Good to have:
  • Experience working on and contributing to Inference serving solutions
  • Knowledge of compiler algorithms for heterogeneous systems
  • Knowledge of open source compiler infrastructure like LLVM or gcc
  • Understanding of low-level kernels
Not hearing back from companies?
Unlock the secrets to a successful job application and accelerate your journey to your next opportunity.

Job Description

Job Description

We are looking for a senior contributor to design, develop and optimize AI frameworks for Inference. In this role, you will work with a cross-geo teams to enhance the inference stack to ensure competitive performance on deep learning inference models with a specific focus on the PyTorch framework.

The roles and responsibilities that you would need to performance may include the following:

  • Design and develop SW techniques for AI frameworks - both HW-agnostic and HW-aware
  • Contribute to enhancing and extending the Inference  and Training capabilities in our Software stack
  • Profile deep learning inference workloads as needed and identify optimization opportunities

Qualifications

  • BTech, MS or PhD in CS or related fields with an overall experience of 10 to 15 years
  • Atleast 2 or 3 years of experience working on Inference frameworks/tools for inference for deep learning models and that have been deployed/used by customers
  • Architecture/Design contributions to Inference systems
  • Detailed understanding of machine learning systems optimization and deployment techniques such as quantization
  • Experience with optimization techniques for deployment of Large Language Models (LLMs)
  • Deep implementation knowledge of transformers and inference specific optimizations
  • Programming skills in Advanced C++, Python and parallel programming skills
  • Ability to debug complex issues in multi-layered SW systems
  • Understanding of SW integration across open source frameworks and internal framework layers
  • Strong understanding of computer architecture
  • Effective communication skills and experience with working in a cross-geo setup

Preferred

  • Experience working on and contributing to Inference serving solutions
  • Knowledge of compiler algorithms for heterogeneous systems
  • Knowledge of open source compiler infrastructure like LLVM or gcc
  • Understanding of low-level kernels

Inside this Business Group

The Data Center & Artificial Intelligence Group (DCAI) is at the heart of Intel’s transformation from a PC company to a company that runs the cloud and billions of smart, connected computing devices. The data center is the underpinning for every data-driven service, from artificial intelligence to 5G to high-performance computing, and DCG delivers the products and technologies—spanning software, processors, storage, I/O, and networking solutions—that fuel cloud, communications, enterprise, and government data centers around the world.

Posting Statement

All qualified applicants will receive consideration for employment without regard to race, color, religion, religious creed, sex, national origin, ancestry, age, physical or mental disability, medical condition, genetic information, military and veteran status, marital status, pregnancy, gender, gender expression, gender identity, sexual orientation, or any other characteristic protected by local law, regulation, or ordinance.

Benefits

We offer a total compensation package that ranks among the best in the industry. It consists of competitive pay, stock, bonuses, as well as, benefit programs which include health, retirement, and vacation. Find more information about all of our Amazing Benefits It has come to our notice that some people have received fake job interview letters ostensibly issued by Intel, inviting them to attend interviews in Intel’s offices for various positions and further requiring them to deposit money to be eligible for the interviews. We wish to bring to your notice that these letters are not issued by Intel or any of its authorized representatives. Hiring at Intel is based purely on merit and Intel does not ask or require candidates to deposit any money. We would urge people interested in working for Intel, to apply directly at https://jobs.intel.com/ and not fall prey to unscrupulous elements.

Working Model

This role will be eligible for our hybrid work model which allows employees to split their time between working on-site at their assigned Intel site and off-site. * Job posting details (such as work model, location or time type) are subject to change...
View Full Job Description

About The Company

Intel’s mission is to shape the future of technology to help create a better future for the entire world. By pushing forward in fields like AI, analytics and cloud-to-edge technology, Intel’s work is at the heart of countless innovations. From major breakthroughs like self-driving cars and rebuilding the coral reefs, to things that make everyday life better like blockbuster effects and improved shopping experiences — they’re all powered by Intel technology. With a career at Intel, you have the opportunity to help make the future more wonderful for everyone.

San José Province, Costa Rica (Hybrid)

San José Province, Costa Rica (On-Site)

San José Province, Costa Rica (On-Site)

Ireland (Hybrid)

Arizona, United States (Hybrid)

San José Province, Costa Rica (Hybrid)

Jalisco, Mexico (Hybrid)

Jalisco, Mexico (Hybrid)

View All Jobs

Level Up Your Career in Game Development!

Transform Your Passion into Profession with Our Comprehensive Courses for Aspiring Game Developers.

Job Common Plug