Applied AI (frameworks) Engineer

at Intel · Industrial · Bangalore, India

Engineer to work on Intel's AI frameworks software stack, focusing on design, development, and optimization of features for AI accelerators and GPUs. This includes ML kernel development, enhancing training and inference capabilities, and contributing to open-source AI frameworks like PyTorch, Tensorflow, and JAX.

What you'd actually do

  1. Design and develop SW features for AI frameworks - both HW-agnostic and HW-aware, especially in ML kernel development.
  2. Enhance and extend the Deep learning training, and Inference capabilities in the Software stack.
  3. Identifying optimization opportunities in the software stack to enhance performance of Deep learning workloads
  4. Participate in discussions with Open-source community, involve in development, adopting upstream and Upstream software.

Skills

Required

  • Advanced C++ (C++ 14/17)
  • Intermediate skills of Python
  • parallel programming
  • developing machine learning kernels (GEMM, Convolution, Flash attention)
  • PyTorch, Tensorflow or JAX
  • Deep Learning models/LLMs for Vision / NLP
  • debug complex issues in multi layered SW systems
  • SW integration in large open-source frameworks
  • computer architecture
  • HW-SW optimization techniques
  • working on frameworks/platforms that have gone to production

Nice to have

  • developing and integrating CUTLASS or Triton based kernels in Large language models (LLMs)
  • compiler algorithms for heterogeneous system
  • Fuser optimizations

What the JD emphasized

  • experience in developing machine learning kernels such as GEMM, Convolution, Flash attention etc
  • In depth and hands on experience in one of the frameworks such as PyTorch, Tensorflow or JAX
  • Experience in working on frameworks/platforms that have gone to production

Other signals

  • optimizing state of the art Software stack for Intel's AI accelerators
  • enabling and optimizing Deep learning training, and Inference capabilities
  • ML kernel development
  • working on frameworks/platforms that have gone to production
Read full job description

Job Details:

Job Description:

We are looking for a dynamic and passionate senior contributor to work in Intel's Data Center and AI group (DCAI). Day-to-day work involves working on Open source AI Frameworks such as PyTorch, Tensorflow, JAX etc. The job role involves design and developing features for Intel' AI frameworks software stack. You will be participating in enabling and optimizing state of the art Software stack for Intel's AI accelerators and next generation GPUs. The roles and responsibilities that you would need to performance may include the following: • Design and develop SW features for AI frameworks - both HW-agnostic and HW-aware, especially in ML kernel development. • Enhance and extend the Deep learning training, and Inference capabilities in the Software stack. • Identifying optimization opportunities in the software stack to enhance performance of Deep learning workloads • Participate in discussions with Open-source community, involve in development, adopting upstream and Upstream software.

Qualifications:

BTech or MS/MTech in CS, ECE or related fields with an overall experience of 6 to 12 years. Proficient in Advanced C++ (C++ 14/17) and Intermediate skills of Python and parallel programming. Experience in developing machine learning kernels such as GEMM, Convolution, Flash attention etc In depth and hands on experience in one of the frameworks such as PyTorch, Tensorflow or JAX Practical knowledge of Deep Learning models/LLMs for Vision / NLP Ability to debug complex issues in multi layered SW systems. Understanding of SW integration in large open-source frameworks. Strong understanding of computer architecture and HW-SW optimization techniques. Experience in working on frameworks/platforms that have gone to production. Effective communication skills and experience with working in a cross-geo teams. Preferable Experience in developing and integrating CUTLASS or Triton based kernels in Large language models (LLMs). Knowledge of compiler algorithms for heterogeneous system and Fuser optimizations.BTech or MS/MTech in CS, ECE or related fields with an overall experience of 6 to 12 years. Proficient in Advanced C++ (C++ 14/17) and Intermediate skills of Python and parallel programming. Experience in developing machine learning kernels such as GEMM, Convolution, Flash attention etc In depth and hands on experience in one of the frameworks such as PyTorch, Tensorflow or JAX Practical knowledge of Deep Learning models/LLMs for Vision / NLP Ability to debug complex issues in multi layered SW systems. Understanding of SW integration in large open-source frameworks. Strong understanding of computer architecture and HW-SW optimization techniques. Experience in working on frameworks/platforms that have gone to production. Effective communication skills and experience with working in a cross-geo teams. Preferable Experience in developing and integrating CUTLASS or Triton based kernels in Large language models (LLMs). Knowledge of compiler algorithms for heterogeneous system and Fuser optimizations.

Job Type:

Experienced Hire

Shift:

Shift 1 (India)

Primary Location:

India, Bangalore

Additional Locations:

Business group:

At the Data Center Group (DCG), we're committed to delivering exceptional products and delighting our customers. We offer both broad-market Xeon-based solutions and custom x86-based products, ensuring tailored innovation for diverse needs across general-purpose compute, web services, HPC, and AI-accelerated systems. Our charter encompasses defining business strategy and roadmaps, product management, developing ecosystems and business opportunities, delivering strong financial performance, and reinvigorating x86 leadership. Join us as we transform the data center segment through workload driven leadership products and close collaboration with our partners.

Posting Statement:

All qualified applicants will receive consideration for employment without regard to race, color, religion, religious creed, sex, national origin, ancestry, age, physical or mental disability, medical condition, genetic information, military and veteran status, marital status, pregnancy, gender, gender expression, gender identity, sexual orientation, or any other characteristic protected by local law, regulation, or ordinance.

Position of Trust

N/A

Work Model for this Role

This role will require an on-site presence. * Job posting details (such as work model, location or time type) are subject to change.

  • ADDITIONAL INFORMATION: Intel is committed to Responsible Business Alliance (RBA) compliance and ethical hiring practices. We do not charge any fees during our hiring process. Candidates should never be required to pay recruitment fees, medical examination fees, or any other charges as a condition of employment. If you are asked to pay any fees during our hiring process, please report this immediately to your recruiter.