Intel

Building

Industrial

HQ
Santa Clara, US
Founded
1968
Size
120,000+
Website
intel.com

Currently tracking 64 active AI roles, up 216% versus the prior 4 weeks. Primary focus: Serve · Engineering. Salary range $122k–$414k (avg $253k).

Hiring
64 / 66
Momentum (4w)
+356 +216%
521 opens last 4w · 165 prior 4w
Salary range · avg $253k
$122k–$414k
USD · disclosed roles only
Tracked since
Feb 3
last role today
Hiring velocityscroll left for older weeks
2 new roles
Oct 6
1 new role
Dec 8
3 new roles
Jan 5
5 new roles
12
1 new role
19
2 new roles
26
6 new roles
Feb 2
6 new roles
9
8 new roles
16
18 new roles
23
22 new roles
Mar 2
38 new roles
9
45 new roles
16
29 new roles
23
37 new roles
30
54 new roles
Apr 6
113 new roles
13
110 new roles
20
151 new roles
27
147 new roles
May 4

Jobs (25)

64 AI · 734 total active
TitleStageFunctionLocationFirst seenAI score
AI Software Engineer Intern
This role focuses on building and optimizing a next-generation LLM inference system, including model optimization, inference runtime, and system-level design. It involves research and engineering to implement and optimize core techniques across the stack from model to kernels to runtime to distributed systems, with a key focus on GPU kernel and runtime optimization for an end-to-end AI rack software system for LLM inference.
ServeEngineeringShanghai, China +22w ago9
AI Software Engineer Intern
This role focuses on building and optimizing a next-generation LLM inference system, including model optimization, inference runtime, and system-level design. It involves research and engineering to implement and optimize core techniques across the stack from model to kernels to runtime to distributed systems, with a key focus on GPU kernel and runtime optimization for an end-to-end AI rack software system for LLM inference.
ServeEngineeringShanghai, China +22w ago9
AI Frameworks Software Engineer – Model Compression Algorithm
Develop Intel Neural Compressor product and related tools, optimize for Intel AI platform (CPU, GPU, AI Accelerator). Research and implement quantization and compression techniques for LLMs and text-to-image/video generation models. Track and explore cutting-edge directions in efficient model deployment and inference/finetuning acceleration.
ServePost-trainEngineeringShanghai, China5w ago8
Software Enabling and Optimization Engineer
This role focuses on optimizing AI software solutions for Intel's AI PC environments by collaborating with customers and ecosystem partners. The engineer will develop, integrate, test, tune, and debug software, leveraging tools like OpenVINO, Llama.CPP, Ollama, LMStudio, and vLLM, to enhance product adoption and differentiation. Key responsibilities include researching and prototyping software, evangelizing Intel's tools, leading pre-enabling efforts, and identifying key workloads for future product designs.
ServeEngineeringShanghai, China +13d ago7
AI Framework Software Intern
Internship role focused on optimizing AI software solutions, including algorithms, frameworks, and architectures for computer vision, machine learning, and deep learning. Responsibilities include researching model quantization and graph transformation, evaluating LLM performance on Intel platforms, analyzing software bottlenecks, and assisting in implementing and tuning AI models for performance and accuracy. The role emphasizes hardware-software integration and collaboration for scalable AI solutions.
ServeEngineeringShanghai, China5d ago7
AI GPU Arch Perf Optimization Intern
This internship focuses on optimizing core GPU compute kernels for AI and numerical workloads, validating GPU IP with AI inference and training workloads, and performing GPU performance profiling and analysis. The role involves hardware/software codesign for next-generation Intel GPU and AI accelerator platforms.
ServeEngineeringShanghai, China +12w ago7
AI GPU Arch Perf Optimization Intern
Intern role focused on optimizing GPU compute kernels for AI workloads and validating GPU IP. Involves performance profiling, analysis, and modeling to improve next-generation Intel GPU and AI accelerator platforms.
ServeEngineeringShanghai, China +12w ago7
AI GPU Arch Perf Optimization Intern
This internship focuses on optimizing core GPU compute kernels for AI and numerical workloads, validating GPU IP with AI inference and training workloads, and performing GPU performance profiling and analysis. The role involves hardware/software codesign for next-generation Intel GPU and AI accelerator platforms.
ServeEngineeringShanghai, China +12w ago7
AI Software Engineer Intern
Intern role focused on optimizing CPU kernels for AI workloads, including LLMs and multimodal models, using Intel architecture features and performance profiling tools. Integrates custom operators into production frameworks.
ServeEngineeringShanghai, China +22w ago7
Efficient AI Solutions Engineering Intern
Internship role focused on developing efficient algorithm solutions for accelerating large AI models and agentic systems, with an emphasis on deployment in resource-constrained computing platforms.
ServeAgentEngineeringBeijing, China2w ago7
AI Compiler and Library Engineer - Intern
AI Compiler and Library Engineer Intern at Intel, contributing to the design, development, and optimization of AI software solutions, including algorithms, frameworks, and architectures. Focuses on implementing and tuning models for performance and accuracy, applied research, and hardware-software integration, with potential involvement in system-level deployment. The role emphasizes learning and skill development through hands-on projects supporting Intel's business goals.
ServeEngineeringShanghai, China3w ago7
AI framework vLLM optimization Intern
AI Software Engineering Intern focused on designing, developing, and optimizing AI software solutions, including algorithms, frameworks, and architectures. Key responsibilities include tuning deep learning models, exploring model compression techniques (quantization, pruning), and conducting applied research for system-level deployment and hardware integration. The role emphasizes practical engineering applications and inference optimization.
ServeEngineeringShanghai, China3w ago7
AI Software Engineer Intern
AI Software Engineer Intern role focused on the design, development, and optimization of AI software solutions, including algorithms, frameworks, and architectures. Responsibilities include implementing and tuning models, applied research, hardware-software integration, and system-level deployment. The role involves learning and applying knowledge in areas like computer vision, machine learning, and deep learning, with a focus on performance and accuracy.
ServePost-trainEngineeringShanghai, China3w ago7
Workload optimization intern
This intern role focuses on optimizing deep learning models and their deployment for Intel GPUs/CPUs. Responsibilities include performance tuning, debugging accuracy and memory issues, developing deployment frameworks (e.g., using vLLM), and creating high-performance kernels. The role involves technical syncs with architects and transforming innovative ideas into production-ready features.
ServeEngineeringShanghai, China3w ago7
AI Software Engineering Graduate Intern
AI Software Engineering Graduate Intern to contribute to the development and optimization of AI workload solutions spanning algorithms, AI software frameworks, system software, and infrastructure. Responsibilities include reproducing AI inference workloads, building workload profiles, identifying system bottlenecks, and performing performance analysis to support hardware/software co design for next generation AI SoC products.
ServeEngineeringBeijing, China +13w ago7
AI Framework Engineering
Intel CAIGC is seeking an AI Framework Engineer to advance AI technologies and software ecosystems, focusing on enabling the AI PC ecosystem and delivering full-stack GPU IP solutions. The role involves research, validation, performance evaluation, and optimization of AI models and frameworks on Intel platforms.
ServeEngineeringShanghai, China6w ago7
AI GPU Arch Perf Optimization Intern
Intern role focused on optimizing GPU compute kernels for AI workloads and validating GPU IP. Involves performance profiling, analysis, and building performance models to understand architecture-level behavior, contributing to hardware/software codesign for next-generation Intel GPUs and AI accelerators.
ServeEngineeringShanghai, China +12w ago5
Cloud and AI System Intern
Research intern focusing on system reliability (RAS) and silent data error characterization and mitigation for AI and general-purpose compute platforms, including heterogeneous systems and large-scale server clusters. Responsibilities include designing and running experiments, analyzing logs, and prototyping detection/diagnosis methods to improve data integrity and platform robustness across the HW/FW/OS/runtime stack.
ServeResearchShanghai, China2w ago5
Cloud and AI System Intern
This internship focuses on supporting the design, deployment, and troubleshooting of Cloud and AI systems and solutions, collaborating with hardware and software teams to optimize for real-world use cases. The role involves applying AI framework knowledge and systems engineering principles to ensure reliability and scalability.
ServeEngineeringShanghai, China2w ago5
AI Performance Engineer Intern
AI Performance Engineer Intern at Intel focused on analyzing silicon chip performance for deep learning, conducting large-scale benchmarks, designing automation tools for data collection and analysis, and researching new architectural features for GPUs, CPUs, and SoCs. The role involves system-level modeling, testing, characterization, and performance-per-watt analysis, with a strong emphasis on understanding deep learning models and frameworks.
ServeEngineeringShanghai, China3w ago5
Research Intern for Supernode Solution
Research Intern focusing on system innovation, cost optimization, and GPU interconnect protocols for disaggregated AI supernode architectures. The role involves exploring architectural innovations, implementing distributed memory pooling, and researching Ethernet-native GPU interconnect protocols for large-scale AI inference and training clusters. Familiarity with RDMA, Mellanox tools, and LLM inference benchmarking methodologies is required.
ServePretrainResearchShanghai, China3w ago5
AI Frameworks Engineer - Intern
Internship role focused on the design, development, and optimization of AI software solutions, including algorithms, frameworks, and architectures. The role involves implementing and tuning models for performance, applied research, and hardware-software integration, with potential for system-level deployment. Familiarity with PyTorch, model profiling, and optimization is a plus.
ServeEngineeringShanghai, China3w ago5
AI Frameworks Engineer - Intern
Internship role focused on the design, development, and optimization of AI software solutions, including algorithms, frameworks, and architectures. The role involves implementing and tuning models for performance, applied research, and hardware-software integration, with potential for system-level deployment. Familiarity with PyTorch, model profiling, and optimization is a plus.
ServeEngineeringShanghai, China3w ago5
Web Runtime Optimization Engineer
This role focuses on optimizing web runtime performance for Intel AI PCs by enabling and optimizing key Chromium components, including Web AI features, on Intel hardware (CPU, GPU, NPU). The engineer will work with internal hardware teams and the Chromium open-source community.
ServeEngineeringShanghai, China3w ago5
Cloud Software Development Engineer
Cloud Software Development Engineer at Intel, focusing on optimizing software stacks for Intel hardware in cloud environments. Specializations include data services, AI/ML, or open-source development. Responsibilities involve designing, developing, validating, and debugging software solutions, with a focus on performance optimization and customer support.
ServeEngineeringShanghai, China +14w ago5