Intel
Building- HQ
- Santa Clara, US
- Founded
- 1968
- Size
- 120,000+
- Website
- intel.com
Currently tracking 64 active AI roles, up 216% versus the prior 4 weeks. Primary focus: Serve · Engineering. Salary range $122k–$414k (avg $253k).
Hiring
64 / 66
Momentum (4w)
↑+356 +216%
521 opens last 4w · 165 prior 4w
Salary range · avg $253k
$122k–$414k
USD · disclosed roles only
Tracked since
Feb 3
last role yesterday
Hiring velocityscroll left for older weeks
Jobs (17)
| Title | Stage | AI score |
|---|---|---|
| AI Software Engineer Intern This role focuses on building and optimizing a next-generation LLM inference system, including model optimization, inference runtime, and system-level design. It involves research and engineering to implement and optimize core techniques across the stack from model to kernels to runtime to distributed systems, with a key focus on GPU kernel and runtime optimization for an end-to-end AI rack software system for LLM inference. | Serve | 9 |
| AI Software Engineer Intern This role focuses on building and optimizing a next-generation LLM inference system, including model optimization, inference runtime, and system-level design. It involves research and engineering to implement and optimize core techniques across the stack from model to kernels to runtime to distributed systems, with a key focus on GPU kernel and runtime optimization for an end-to-end AI rack software system for LLM inference. | Serve | 9 |
| AI Frameworks Software Engineer – Model Compression Algorithm Develop Intel Neural Compressor product and related tools, optimize for Intel AI platform (CPU, GPU, AI Accelerator). Research and implement quantization and compression techniques for LLMs and text-to-image/video generation models. Track and explore cutting-edge directions in efficient model deployment and inference/finetuning acceleration. | ServePost-train | 8 |
| Software Enabling and Optimization Engineer This role focuses on optimizing AI software solutions for Intel's AI PC environments by collaborating with customers and ecosystem partners. The engineer will develop, integrate, test, tune, and debug software, leveraging tools like OpenVINO, Llama.CPP, Ollama, LMStudio, and vLLM, to enhance product adoption and differentiation. Key responsibilities include researching and prototyping software, evangelizing Intel's tools, leading pre-enabling efforts, and identifying key workloads for future product designs. | Serve | 7 |
| AI Software Engineering Intern AI Software Engineering Intern role focused on designing, developing, and optimizing AI algorithms and frameworks, with contributions to implementation, tuning, applied research, and prototyping for scalable AI solutions. The role involves working with computer vision, machine learning, and deep learning, and requires programming ability in Python and familiarity with ML frameworks. | ServePost-train | 7 |
| AI Framework Software Intern Internship role focused on optimizing AI software solutions, including algorithms, frameworks, and architectures for computer vision, machine learning, and deep learning. Responsibilities include researching model quantization and graph transformation, evaluating LLM performance on Intel platforms, analyzing software bottlenecks, and assisting in implementing and tuning AI models for performance and accuracy. The role emphasizes hardware-software integration and collaboration for scalable AI solutions. | Serve | 7 |
| AI GPU Arch Perf Optimization Intern This internship focuses on optimizing core GPU compute kernels for AI and numerical workloads, validating GPU IP with AI inference and training workloads, and performing GPU performance profiling and analysis. The role involves hardware/software codesign for next-generation Intel GPU and AI accelerator platforms. | Serve | 7 |
| AI GPU Arch Perf Optimization Intern Intern role focused on optimizing GPU compute kernels for AI workloads and validating GPU IP. Involves performance profiling, analysis, and modeling to improve next-generation Intel GPU and AI accelerator platforms. | Serve | 7 |
| AI GPU Arch Perf Optimization Intern This internship focuses on optimizing core GPU compute kernels for AI and numerical workloads, validating GPU IP with AI inference and training workloads, and performing GPU performance profiling and analysis. The role involves hardware/software codesign for next-generation Intel GPU and AI accelerator platforms. | Serve | 7 |
| AI Software Engineer Intern Intern role focused on optimizing CPU kernels for AI workloads, including LLMs and multimodal models, using Intel architecture features and performance profiling tools. Integrates custom operators into production frameworks. | Serve | 7 |
| Efficient AI Solutions Engineering Intern Internship role focused on developing efficient algorithm solutions for accelerating large AI models and agentic systems, with an emphasis on deployment in resource-constrained computing platforms. | ServeAgent | 7 |
| AI Compiler and Library Engineer - Intern AI Compiler and Library Engineer Intern at Intel, contributing to the design, development, and optimization of AI software solutions, including algorithms, frameworks, and architectures. Focuses on implementing and tuning models for performance and accuracy, applied research, and hardware-software integration, with potential involvement in system-level deployment. The role emphasizes learning and skill development through hands-on projects supporting Intel's business goals. | Serve | 7 |
| AI framework vLLM optimization Intern AI Software Engineering Intern focused on designing, developing, and optimizing AI software solutions, including algorithms, frameworks, and architectures. Key responsibilities include tuning deep learning models, exploring model compression techniques (quantization, pruning), and conducting applied research for system-level deployment and hardware integration. The role emphasizes practical engineering applications and inference optimization. | Serve | 7 |
| AI Software Engineer Intern AI Software Engineer Intern role focused on the design, development, and optimization of AI software solutions, including algorithms, frameworks, and architectures. Responsibilities include implementing and tuning models, applied research, hardware-software integration, and system-level deployment. The role involves learning and applying knowledge in areas like computer vision, machine learning, and deep learning, with a focus on performance and accuracy. | ServePost-train | 7 |
| Workload optimization intern This intern role focuses on optimizing deep learning models and their deployment for Intel GPUs/CPUs. Responsibilities include performance tuning, debugging accuracy and memory issues, developing deployment frameworks (e.g., using vLLM), and creating high-performance kernels. The role involves technical syncs with architects and transforming innovative ideas into production-ready features. | Serve | 7 |
| AI Software Engineering Graduate Intern AI Software Engineering Graduate Intern to contribute to the development and optimization of AI workload solutions spanning algorithms, AI software frameworks, system software, and infrastructure. Responsibilities include reproducing AI inference workloads, building workload profiles, identifying system bottlenecks, and performing performance analysis to support hardware/software co design for next generation AI SoC products. | Serve | 7 |
| AI Framework Engineering Intel CAIGC is seeking an AI Framework Engineer to advance AI technologies and software ecosystems, focusing on enabling the AI PC ecosystem and delivering full-stack GPU IP solutions. The role involves research, validation, performance evaluation, and optimization of AI models and frameworks on Intel platforms. | Serve | 7 |