Currently tracking 20 active AI roles, with 14 new openings in the last 4 weeks. Primary focus: Serve · Engineering. Salary range $160k–$300k (avg $226k).
Data AI · Open-source model infra
| Title | Stage | AI score |
|---|---|---|
| Research Engineer, Core ML Research Engineer role focused on improving inference efficiency and unifying it with RL/post-training systems for production-grade AI APIs. The role involves end-to-end ownership of critical systems, translating frontier ideas into robust infrastructure, and shipping measurable improvements in latency, throughput, cost, and model quality at scale. | ServePost-train | 10 |
| Forward Deployed Engineer (Inference & Post-Training) Forward Deployed Engineer focused on optimizing inference engines and fine-tuning pipelines for production AI teams, acting as a technical partner to strategic customers. Responsibilities include inference engine optimization, performance tuning, post-training/fine-tuning (LoRA, SFT, DPO, RLHF, GRPO), customer alignment, onboarding, and providing product feedback. | ServePost-train | 9 |
| Senior Machine Learning Engineer, Voice AI Senior ML Engineer focused on optimizing the model serving layer for voice AI workloads, including speech-to-text and text-to-speech models. The role involves hands-on work with inference engines, GPU optimization, batching strategies, and ensuring new model architectures can be productionized efficiently. The goal is to achieve best-in-class latency and reliability for real-time voice applications. | Serve | 9 |
| Research Engineer, Frontier Speculative Decoding Research Engineer focused on translating internal model training research into production-ready deployments by fine-tuning general-purpose models into specialized tools. This involves designing novel speculative algorithms, data curation, hyperparameter tuning, and checkpoint evaluation, with a focus on accuracy-efficiency tradeoffs for generative AI models. | Post-trainServe | 9 |
| Systems Research Engineer, GPU Programming This role focuses on optimizing and developing GPU-accelerated kernels and algorithms for ML/AI applications, requiring expertise in GPU programming (CUDA, Triton) and performance profiling. The engineer will collaborate with modeling, hardware, and software teams to enhance AI system efficiency and co-design GPU architectures. | Serve | 9 |
| AI Researcher, Core ML (Turbo) AI Researcher focused on the intersection of efficient inference algorithms, architectures, engines, and post-training/RL systems for production-scale API services. The role involves advancing inference efficiency, unifying inference with RL/post-training, and owning critical systems. | ServePost-train | 9 |
| Forward Deployed Engineer (GPU Clusters) The Forward Deployed Engineer (FDE) will be a technical partner to customers building large-scale AI models, focusing on GPU cluster infrastructure, networking, storage, and orchestration to ensure stability, optimize performance, and facilitate platform adoption. This role involves hardening clusters, tuning orchestration layers (Kubernetes/SLURM), debugging low-level bottlenecks, building reference designs, and leading benchmarking exercises. | Serve | 8 |
| Engineering Manager, Model Serving Engineering Manager for Together AI's Model Serving platform, focusing on delivering world-class inference and fine-tuning in public APIs and customer deployments. Responsibilities include owning SLAs, improving testing/deployment/monitoring, building self-serve tooling, defining configuration best practices for inference engines, leading incident response, and mentoring team members. Requires 5+ years operating production ML inference or training systems at scale and 2+ years in senior IC or tech lead roles, with deep expertise in Kubernetes, multi-cluster orchestration, and ML serving frameworks. | ServePost-train | 8 |
| Machine Learning Engineer Machine Learning Engineer at Together AI focused on developing and scaling production systems for LLM inference and fine-tuning APIs. Requires strong experience in high-performance, distributed systems and the LLM inference ecosystem. | ServePost-train | 8 |
| Machine Learning Engineer - Inference Machine Learning Engineer focused on optimizing and enhancing the performance of AI inference systems, working with state-of-the-art large language models to ensure efficient and effective operation at scale. Responsibilities include designing and building production systems, optimizing runtime inference services, and creating supporting tools and documentation. | Serve | 8 |
| Senior Platform Engineer, Voice AI Senior Platform Engineer for Together AI's Voice AI platform, focusing on the API and infrastructure layer for real-time speech-to-text and text-to-speech models. The role involves building WebSocket and HTTP APIs, designing autoscaling for latency-sensitive streaming, and ensuring platform reliability for production voice agents. | Serve | 7 |
| Senior Backend Engineer, Inference Platform Senior Backend Engineer focused on building and optimizing the inference platform for advanced generative AI models, including LLMs and multimodal models, at scale. The role involves optimizing latency, throughput, and resource allocation across tens of thousands of GPUs, collaborating with researchers to productionize frontier models, and contributing to open-source inference projects. | Serve | 7 |
| Machine Learning, Platform Engineer Machine Learning Platform Engineer at Together AI, focusing on building a container platform, optimizing autoscaling, minimizing cold starts, and improving end-to-end model performance for custom models and dedicated inference. The role involves optimizing inference across the stack, including CUDA kernels, PyTorch, inference engines, and container orchestration. | Serve | 7 |
| AI Infrastructure Engineer AI Infrastructure Engineer responsible for keeping user-facing services and production systems running smoothly, applying engineering principles and automation to operating environments. Focuses on systems, availability, reliability, and scalability, with interests in algorithms and distributed systems. Builds and runs infrastructure using Ansible, Terraform, and Kubernetes, and designs monitoring systems. | Serve | 7 |
| Senior Software Engineer - Together Cloud Infrastructure Senior Software Engineer focused on building and operating a high-performance, global AI cloud infrastructure platform. This includes designing and maintaining backend services for hardware management, IaaS software layer for GPU data centers, high-performance object storage for pretraining datasets, and advanced observability stacks for distributed pretraining. The role also involves architecture and research for decentralized AI workloads and contributing to the open-source platform. | ServeData | 7 |
| Solutions Architect Solutions Architect at Together AI to work with customers and prospects to create business value through Generative AI applications. This role involves acting as a technical advisor, running demonstrations and POCs, collaborating with sales, building relationships with customer leadership, delivering feedback to product/engineering/research, and building educational content. Requires 5+ years in a customer-facing technical role with 2+ years in pre-sales, strong technical background in AI/ML/GPU, understanding of LLM training/fine-tuning/inference, Python/JavaScript proficiency, and familiarity with infrastructure services. | Serve | 7 |
| Staff Engineer, Customer Insights Staff Engineer to build and scale the customer-facing visibility layer for Together's AI Cloud, focusing on historical analytics, activity history, audit logs, event timelines, notifications, and investigation workflows. The role will evolve these foundations into AI-first investigation and insight workflows that summarize activity, explain anomalies, and provide trustworthy context for human operators and autonomous agents. This is a hands-on role designing event, query, delivery, and governance systems, and building user-facing workflows for enterprise customers. | — | 5 |
| Technical Account Manager (TAM), AI Factory This role is a Technical Account Manager focused on the infrastructure supporting large-scale AI GPU deployments for a strategic enterprise customer. The TAM will be the primary technical point of contact, responsible for the end-to-end technical relationship across compute, networking, storage, and facilities, ensuring smooth delivery and operational health. Responsibilities include issue lifecycle management, hardware lifecycle management, advising on infrastructure stack best practices, owning the observability strategy, coordinating operations, and managing capacity expansions. The role requires deep expertise in GPU infrastructure, large-scale networking, enterprise storage, and DC operations, with experience in customer-facing technical roles and AI/HPC infrastructure. | — | 5 |
| Director, Support Engineering This role leads and scales the customer support function for Together AI, focusing on both API support (serverless/dedicated inference, billing) and GPU support (large-scale training infrastructure). It's a player-coach position requiring hands-on involvement in complex escalations, managing support engineers, defining KPIs, and improving support workflows and tooling. The role requires strong technical depth in AI infrastructure, distributed systems, and experience with SLA-driven operations. | — | 5 |
| Customer Support Engineer (GPU Cluster) Customer Support Engineer role focused on supporting customers using Together AI's GPU clusters for training, fine-tuning, and inference. The role involves resolving complex technical challenges, acting as a product expert, and collaborating with Engineering and Product teams. Requires experience in customer-facing technical roles, familiarity with AI/ML, GPU technologies, and infrastructure services like Kubernetes. | — | 5 |
| Sr. Partnerships Manager, Model Ecosystem This role is responsible for building and managing the model ecosystem for Together AI, focusing on negotiating deals with model builders to bring proprietary and open-source models onto the platform. It involves working closely with Product, Finance, and Marketing to ensure the model roadmap is technically superior, commercially viable, and market-facing. The role requires strong deal-making, technical curiosity, and experience in business development or strategic partnerships within developer platforms. | — | 5 |
| Backend Software Engineer — Data Platform & AI Data Products Backend Software Engineer focused on building data platform infrastructure and LLM-adjacent data products. The role involves designing and developing backend services for event streams, access layers, and APIs, as well as creating services for prompt categorization, enrichment, and metadata. The engineer will apply AI augmentation mindset to their own development and the systems they build, with a focus on production backend systems, distributed systems, and data modeling. | Serve | 5 |
| Lead Product Designer Lead Product Designer to craft user experiences for technical AI development tools, shape AI development, and establish design standards for a growing organization. This role involves leading UX initiatives, elevating design quality, and collaborating with Engineering, Product, and Marketing. | — | 5 |
| Product Marketing Director Product Marketing Director at Together AI, a frontier AI cloud company. This role will own platform and product value propositions, GTM strategy, product launches, and messaging. The role involves leading and scaling the PMM function, partnering with Product Management, Sales, and Engineering. Requires 10+ years of PMM experience in enterprise software (preferably AI/Cloud) and 5+ years in team leadership. The company has seen significant growth and is research-driven, contributing to open-source AI advancements. | — | 5 |
| Senior Software Engineer - Together Cloud Platform Senior Backend Engineer role focused on building and scaling the AI Acceleration Cloud platform, which virtualizes ML hardware and provides self-serve AI cloud services for ML practitioners. Responsibilities include developing distributed GPU scheduling systems, global management planes, and customer-facing cloud platform services, ensuring high availability and performance. | — | 5 |
| Sr. Technical Program Manager (TPM) This role is for a Senior Technical Program Manager (TPM) at an AI infrastructure company. The TPM will focus on building, optimizing, and scaling global GPU resources, ensuring efficient and reliable operation of the AI model backbone. Responsibilities include product development for AI researchers and customers, owning the product roadmap, stakeholder engagement, and cross-functional execution across Research, Engineering, DevOps, SRE, and Go-to-Market teams. Requires 5+ years of experience in AI/ML product or infrastructure, with a technical background. | — | 5 |
| Strategic Finance Senior Manager Strategic Finance Senior Manager at Together AI, a research-driven AI infrastructure company. This role focuses on providing financial insights, driving strategic decision-making, and optimizing business performance, with a significant emphasis on guiding the optimization and scaling of the company's compute infrastructure. The position requires strong financial modeling, business judgment, and the ability to partner with various functions like Engineering, Product, and GTM. | — | 5 |
| Finance Analytics Engineer This role is for a Finance Analytics Engineer who will own the data layer for the Finance team, building models, pipelines, and reporting infrastructure. Responsibilities include owning the dbt transformation layer, orchestrating runs with Airflow, delivering dashboards, partnering with finance teams, setting data quality standards, and building a data foundation to support AI automation. Requires 5+ years of experience in analytics engineering or data engineering, with expertise in SQL, dbt, Snowflake, and Airflow, and strong dimensional modeling fundamentals. | — | 0 |
| Staff Backend Engineer - Commerce Staff Backend Engineer to own the technical vision, architecture, and execution of the commerce platform powering Together's Cloud products, including usage-based billing, payment processing, customer-facing analytics, and product entitlements. This role requires deep expertise in backend systems, distributed systems, and API design, with a focus on scalability, fault tolerance, and influencing cross-functional teams. | — | 0 |
| Director, Data Center Operations This role is for a Director of Data Center Operations at Together AI, focusing on building and scaling the physical infrastructure for AI workloads. The responsibilities include designing and commissioning data center white space, managing power and cooling systems, and building a break-fix team. It is a ground-floor, builder role with ownership over operational foundations. | — | 0 |
| Executive Assistant- Finance and Infrastructure Strategy Executive Assistant supporting Finance and Infrastructure Strategy leadership at an AI company. Responsibilities include managing schedules, travel, expenses, and contracts. Requires 4-6+ years of experience supporting senior leadership, strong organizational and problem-solving skills. | — | 0 |
| Analytics Engineer — Data Warehouse Staff Analytics Engineer role focused on building and maintaining the data warehouse transformation layer using dbt and Airflow. The role involves dimensional modeling, data quality, governance, and stakeholder management, with a focus on financial and billing data. The company is an AI infrastructure and platform company. | — | 0 |
| Director of Tax Seeking a Director of Tax to establish and scale the tax function from the ground up, defining global tax strategy and managing compliance across multiple jurisdictions for a high-growth AI infrastructure company. | — | 0 |
| Staff Engineer, Product UI Platform Staff Engineer to own and evolve the Product UI Platform, the architectural foundation for full-stack features across the web surface. This role will drive the technical direction of the Next.js/typescript/nodejs web runtime, BFF layer, and application integration patterns, evolving the product runtime from a monolithic growth architecture to a scalable, modular, and high-leverage platform. | — | 0 |
| Data Warehouse Engineer Staff Data Warehouse Engineer responsible for designing, operating, and evolving a data warehouse stack (bronze/silver/gold), owning core data models and metrics, and establishing data quality and governance standards. The role involves building and maintaining data pipelines, designing analytics-ready models, leading Master Data Management patterns, implementing data quality checks, and building a business semantic layer. The engineer will use SQL, Python, and Spark, mentor junior engineers, and contribute to technical standards. | — | 0 |
| Sales and Marketing Operations Manager This role is for a Sales and Marketing Operations Manager at Together AI, a research-driven AI company. The manager will optimize the lead-to-revenue funnel, develop reporting dashboards, analyze performance across the tech stack, manage attribution and ROI, assist with forecasting, oversee the tech stack (Salesforce, marketing automation, sales engagement tools), implement workflows for lead routing and scoring, ensure data integrity, and align Sales and Marketing strategies. Requirements include a Bachelor's degree, 5-7 years of experience in Sales/Marketing Ops or BI, expertise in marketing automation and CRMs, data visualization skills, strong analytical abilities, and advanced Excel skills (SQL, R, or Python are a plus). | — | 0 |
| Staff Engineer, API Core Platform Staff Engineer to found the API Platform team, focusing on building and scaling core systems and architecture for Together AI's mission-critical APIs. Responsibilities include improving the backend API layer, designing next-gen API platform solutions, and ensuring reliability, performance, and consistency across public and client APIs. The role requires deep hands-on experience with critical-path code and building platforms that unify engineering efforts. | — | 0 |
| Payroll Manager Seeking a Payroll Manager to lead and scale payroll operations, ensuring accurate, compliant, and timely processing for all employees. Responsibilities include managing full-cycle payroll, overseeing bi-weekly processing, reviewing payroll inputs, serving as a subject matter expert on regulations, managing visa-related payroll, and handling incentive/equity compensation. Requires 6+ years of hands-on payroll experience, strong knowledge of payroll regulations, and experience with complex multi-state payrolls and equity compensation. | — | 0 |
| Senior Technical Recruiter This role is for a Senior Technical Recruiter at Together AI, a company building an AI Acceleration Cloud. The recruiter will partner with engineering leaders to drive hiring for core engineering functions, manage the candidate journey, provide market intelligence, and design interview processes. The company focuses on the generative AI lifecycle, AI cloud infrastructure, and open-source AI research. | — | 0 |
| Senior Developer Productivity Engineer Senior Developer Productivity Engineer at Together AI, a research-driven AI company. Focuses on optimizing engineering workflows, CI/CD pipelines, and building shared tooling to accelerate software delivery. Requires strong experience in DevOps, CI/CD, and Python/Go/TypeScript. | — | 0 |
| Senior Data Engineer Senior Data Engineer to build and operate data infrastructure for billing, analytics, and BI tools. Requires expertise in stream processing, real-time analytics, and IaC. Role involves designing, building, and scaling data platforms in a fast-paced environment. | — | 0 |
| Senior Network Engineer Senior Network Engineer responsible for designing, implementing, and maintaining network infrastructure for AI company's user-facing services and production systems. Focus on routing, switching, network security, and protocols, with an emphasis on automation and HPC-based data center networking. Experience with large-scale hybrid data center networks, TCP/IP, BGP, OSPF, VXLAN, EVPN, QoS, and network automation tools (Python, Ansible). Proficient in network troubleshooting tools and Linux environments. Experience with cloud networks (AWS, GCP, Azure) and multi-vendor network devices (Cisco, Arista, Juniper, Mellanox). Preferred knowledge of RoCE, Infiniband, Docker, Kubernetes, Slurm, and AI training workloads. | — | 0 |