1 - 20 of 37,733 Jobs

Search for AI Infrastructure Specialist

37,733 results

Apply
vcluster logovcluster logo
Full-time|Remote|Remote- US

Join vCluster as an AI Infrastructure Specialist, where you will engage directly with clients at a pivotal stage in their journey—from configuring bare metal GPU nodes to deploying production-ready solutions. This role transcends typical professional services; you will operate in a pre-sales capacity, focusing on proof of value engagements that lead to robus…

Mar 30, 2026
Apply
Savvy logoSavvy logo
Full-time|On-site|NYC Office

Join our innovative team at Savvy as an Applied AI Infrastructure Specialist. In this role, you will leverage cutting-edge technologies to enhance our AI capabilities and streamline infrastructure processes. Collaborate with a dynamic team of experts to implement solutions that drive efficiency and effectiveness within our systems. This position offers an exciting opportunity to work at the forefront of artificial intelligence and infrastructure development.

Mar 10, 2026
Apply
Hammerhead AI logo
Full-time|On-site|Redwood City

Hammerhead AI develops orchestration solutions that enhance power efficiency for artificial intelligence. The company’s platform upgrades data center power infrastructure, allowing AI token generation to be optimized within current electrical limits. This method helps avoid building new power plants or expanding the grid. Hammerhead AI has optimized more than 8 gigawatts of mission-critical power across the globe, serving a $64 billion annual market and working to reduce the environmental footprint of AI infrastructure. Role overview The Business Development Lead plays a key part in driving adoption of Hammerhead AI’s compute solutions. This is a hands-on position that shapes commercial strategy, influences how the company is positioned in the market, and supports early revenue growth. The role reports directly to the Head of Go-To-Market (GTM). What you will do Develop and refine Hammerhead AI’s commercial strategy, building a prioritized go-to-market plan. Manage the sales pipeline from initial contact through technical discovery, solution design, proposal, and closing. Monitor trends in AI compute demand, GPU supply, power constraints, and data center deployment models, turning these insights into actionable strategies. Location This position is based in Redwood City.

Apr 28, 2026
Apply
Genesis Molecular AI logo
Full-time|On-site|NYC or SF Bay Area

Genesis Molecular AI is building the GEMS molecular AI platform, driving advances in foundation model training and industrial screening. Strategic partnerships and a strong compute infrastructure are central to the company’s growth and mission. Role Overview The Director of AI Infrastructure Partnerships will lead efforts to secure and manage critical technology alliances, investments, and compute resources. This leader will work closely with top AI organizations, hardware providers, and investors, including firms like a16z and NVIDIA, to support Genesis’s technical and business goals. The role is based in either New York City or the San Francisco Bay Area. What You Will Do Oversee partnerships with NVIDIA and identify new opportunities with leading AI organizations. Structure contracts, equity deals, technical collaborations, co-publications, and data-sharing agreements for both public and proprietary experimental and synthetic data. Create presentations and written materials that clearly communicate Genesis’s platform vision and technical strengths to partners and investors, and integrate these messages into broader external communications. Serve as the business lead and chief negotiator for major cloud computing and AI infrastructure deals. Secure high-performance compute at competitive rates and maintain strong relationships with key partners. Monitor the AI compute market, evaluating providers for cost, reliability, and availability to support research and deployment needs. Work with ML Engineering to forecast compute requirements for model training, synthetic data generation, fine-tuning, and large-scale inference. Optimize performance and budget across multiple cloud environments and track usage to maximize value. Manage the internal budgeting process for compute spend. Translate technical needs into financial forecasts and present capital allocation recommendations to company leadership. What We’re Looking For Significant experience in AI and cloud computing, including managing high-value negotiations and partnerships. Strong analytical and strategic skills, with the ability to assess market trends and make informed decisions. Excellent communication and interpersonal abilities, comfortable explaining complex topics to a range of audiences.

Apr 15, 2026
Apply
Lavendo logoLavendo logo
Full-time|$225K/yr - $315K/yr|Remote|San Francisco

About the CompanyLavendo is a pioneering publicly traded company leading the charge in the AI revolution. With an AI-centric cloud platform, we are transforming the artificial intelligence landscape. Our state-of-the-art infrastructure, including extensive GPU clusters and advanced cloud services, supports developers in harnessing the explosive growth of the global AI industry, catering to Fortune 1000 firms, innovative startups, and AI researchers alike.Company type: Publicly tradedIndustry: AI/ML, Cloud Computing, Infrastructure-as-CodeCandidate Location: Remote U.S.Our mission is to democratize AI infrastructure access and empower organizations to innovate, optimize, and deploy AI solutions seamlessly at any scale. By simplifying the complexities of AI development, we provide a comprehensive full-stack AI platform that marries robust hardware with easy-to-use tools and services.The OpportunityWe are on the lookout for a Senior AI/ML Specialist Solutions Architect to become a crucial part of our client's dynamic team. This role presents an exciting opportunity to design and implement scalable AI solutions tailored for AI-centric clients, leveraging cutting-edge technologies and contributing to one of the most powerful commercially available supercomputers.What You'll DoArchitect and enhance distributed training and inference systems for large-scale AI models.Design and deliver customer-centric solutions that optimize performance and drive business value.Lead the migration of ML pipelines from Proof of Concept to scalable production environments.Foster long-term relationships with clients, ensuring satisfaction and alignment with their strategic objectives.Produce whitepapers, conduct technical presentations, and facilitate webinars to disseminate insights and best practices.Provide technical guidance and mentorship to teams regarding AI infrastructure and deployment strategies.Collaborate with engineering and product teams to prioritize customer feedback and shape product roadmaps.

Feb 23, 2026
Apply
Hammerhead AI logo
Full-time|On-site|Redwood City

Hammerhead AI develops software that advances AI infrastructure by optimizing how data centers use power. The company’s platform manages power, cooling, and compute resources in real time, using reinforcement learning to help operators run more AI workloads without exceeding electrical or thermal limits. So far, Hammerhead AI has optimized over 8 gigawatts of critical power worldwide, reducing the environmental impact of large-scale computing. This team brings together expertise in AI, energy, and computing. Employees collaborate with specialists in reinforcement learning, IoT, and infrastructure, working toward more efficient and sustainable AI operations. The company values innovation in data center design and offers competitive compensation, equity, and benefits. Many team members have experience building and scaling successful startups. Role overview The Product Manager for the AI Infrastructure Platform leads the vision, roadmap, and execution for Hammerhead’s orchestration platform. This role defines product offerings, clarifies their purpose, and plans delivery for customers. The work involves translating technical details about data center power, cooling, and compute systems into actionable requirements for engineering and deployment teams. What you will do Own and communicate the product vision and roadmap for the AI Infrastructure Platform Define product features and clarify their value to customers Translate technical knowledge of power, cooling, and compute systems into clear requirements Work closely with reinforcement learning engineers, software teams, deployment, customer success, and go-to-market groups Ensure the platform delivers measurable value at every customer site What sets this role apart Work at the intersection of AI, energy, and computing Collaborate with experts in reinforcement learning and infrastructure Contribute to a more sustainable foundation for AI computing Join a team driving innovation in data center operations

Apr 28, 2026
Apply
novita-ai logo
Full-time|On-site|San Francisco

About Us:At novita-ai, we are a rapidly growing global provider of AI cloud infrastructure, leading the charge in the artificial intelligence revolution. Our innovative platform equips developers and enterprises with powerful, scalable, and user-friendly solutions such as Model APIs, GPU Instances, and Serverless Computing. As organizations around the globe strive to integrate AI into their offerings, we serve as the essential engine that fuels their innovative efforts.Join our world-class team and contribute to our expanding customer base. This unique opportunity allows you to be part of a dynamic company in a hyper-growth market, where your technical skills will directly impact customer success and drive our business forward.The Role:As a Solutions Engineer, you will act as the primary technical leader and trusted advisor for our clients throughout their journey. You will collaborate closely with the sales team to bridge the gap between complex customer challenges and our sophisticated technical solutions. Your mission is to build technical credibility, demonstrate the capabilities of our platform, and design tailored solutions that empower our clients to achieve their AI-related business objectives.What You'll Do:Technical Discovery & Solution Design: Collaborate with Account Executives to gain a deep understanding of customer needs, technical requirements, and business goals. Develop elegant and effective solutions utilizing our AI infrastructure stack (Model APIs, GPU Instances, Serverless).Product Demonstration & Proof of Concept (POC): Conduct engaging, customized product demonstrations and interactive workshops. Plan, manage, and execute successful POCs, showcasing the value and performance of our platform within the client’s environment.Technical Evangelism & Trusted Advisory: Communicate the value proposition of our platform to diverse audiences, including both technical and non-technical stakeholders, from engineers to C-level executives. Establish yourself as the go-to expert for customers on best practices in AI infrastructure.Sales Enablement & Market Feedback Loop: Create and maintain technical sales materials, including whitepapers, best practice guides, and demo scripts. Serve as the voice of the customer, relaying valuable feedback from the field to our Product and Engineering teams to influence our product roadmap.Onboarding & Implementation Guidance: Facilitate a seamless post-sales transition by providing initial onboarding support and architectural guidance, setting customers up for sustained success.

Aug 27, 2025
Apply
Reflection AI logoReflection AI logo
Full-time|On-site|San Francisco

Reflection AI builds open weight models for a wide range of users, including individuals, businesses, and governments. The team brings together talent from organizations such as DeepMind, OpenAI, Google Brain, Meta, Character.AI, and Anthropic, all working to advance open superintelligence. Role overview The AI Compute and Infrastructure Counsel acts as the main legal advisor to Reflection AI’s Strategy and Operations teams on complex infrastructure initiatives. Based in San Francisco, this attorney leads negotiations and manages agreements that support the company’s growing AI infrastructure. The work spans collaborations with hardware manufacturers, cloud capacity deals, and contracts related to data centers, utilities, and new facility builds. This position is designed for a commercial lawyer with experience at the intersection of advanced AI and infrastructure. The role provides autonomy, the opportunity to establish legal frameworks for a new function, and a direct impact on the company’s AI systems. What you will do Negotiate compute and cloud capacity agreements with hyperscalers, neoclouds, and new vendors, covering terms like capacity reservations, service-level commitments, portability, and exit rights. Manage hardware partnerships with vendors in chips, accelerators, servers, and networking. Oversee legal support for data center and AI facility projects, including master agreements for colocation and hosting, ground leases, build-to-suit leases, construction contracts, interconnection agreements, and power purchase agreements. Structure and negotiate power arrangements, such as power purchase agreements, tolling agreements, utility service contracts, behind-the-meter generation, and long-term energy deals. Lead legal work on strategic infrastructure transactions, including joint ventures, site acquisitions, and custom financing models for the AI factory roadmap. Develop scalable playbooks, templates, and delegation systems to help commercial and infrastructure teams operate efficiently and maintain high standards. Collaborate with Security, Privacy, and Policy teams on matters like tenant isolation, customer data handling, and sovereign compute requirements.

Apr 28, 2026
Apply
Andromeda logoAndromeda logo
Full-time|Remote|Global Remote / San Francisco, CA

Are you passionate about artificial intelligence and have extensive experience in AI infrastructure? Although there may not be a specific position that matches your qualifications at this moment, we encourage you to submit your resume here. Should a suitable opportunity arise in the future that aligns with your expertise, we will be sure to connect with you.

Feb 26, 2026
Apply
The Allen Institute logoThe Allen Institute logo
Full-time|On-site|Seattle, WA

The Allen Institute is seeking a visionary and experienced Director of AI Infrastructure to lead our innovative efforts in artificial intelligence. In this pivotal role, you will oversee the development and implementation of robust AI systems that drive our scientific endeavors.As the Director of AI Infrastructure, you will work closely with cross-functional teams to ensure the seamless integration of AI technologies into our research processes. You will be responsible for managing a talented team of engineers and data scientists, fostering a culture of innovation and collaboration.

Mar 25, 2026
Apply
Jobgether logoJobgether logo
Full-time|Remote|US

Role overview This fully remote Senior Infrastructure Engineer - AI Platform Specialist position focuses on building and improving the core infrastructure for a large-scale AI platform. The platform supports machine learning workflows across several industries, including autonomous vehicles, healthcare, and agriculture. Key responsibilities Design and enhance infrastructure for both individual researchers and enterprise deployments. Develop scalable systems to manage unstructured data efficiently. Maintain high standards for reliability, security, and performance throughout the data platform. Support enterprise production environments, balancing technical work with customer-facing tasks. Work closely with engineering teams to improve developer productivity and advance infrastructure maturity. Impact Work in this role directly shapes the scalability and reliability of AI systems used worldwide. Contributions influence the evolution and performance of mission-critical data platforms across multiple sectors.

Apr 28, 2026
Apply
Retell AI logo
Full-time|On-site|San Francisco Bay Area

Join the Revolution at Retell AIRetell AI is pioneering the future of call centers through innovative voice AI, driven by first principles thinking.In just 18 months since our inception, we have empowered thousands of businesses with our AI voice agents, transforming how sales, support, and logistics calls are managed—previously requiring extensive human teams. Supported by prestigious investors such as Y Combinator and Alt Capital, we've rapidly scaled from $5M ARR to an impressive $36M ARR with a compact yet dynamic team of 20.Our ambition for 2026 is to create a revolutionary customer experience platform, where entire contact centers are powered by AI. Moving beyond basic automation, we aim to develop intelligent AI “workers” that serve as frontline agents, QA analysts, and managers, continuously enhancing customer interactions without the need for constant human oversight.As we expand, we are seeking passionate engineers who are eager to solve challenging technical problems, act swiftly, and make a significant impact in one of the fastest-growing voice AI startups. Let’s shape the future together.

Aug 12, 2025
Apply
Laurel logoLaurel logo
Full-time|Remote|Remote, Latin America

Laurel builds AI-powered solutions that help professional service firms reclaim and optimize their most valuable resource: time. Our platform automates time capture and connects time data to business results, supporting firms like EY, Aprio, Crowell & Moring, and Frost Brown Todd. Each year, we process over 1 billion work activities, providing insights that were previously out of reach. Our team brings together expertise in AI, product development, and engineering. We focus on practical innovation, aiming to double the value professionals get from their time. We value humility, curiosity, and a drive to improve how knowledge work happens. This remote role is based in Latin America and centers on managing Laurel’s internal IT and AI infrastructure. The position guides the Business Technology team, supports employees, and ensures that our systems meet the high security standards required by our clients. As the team grows, this manager will play a key role in shaping its direction and capabilities. What you will do IT Operations and Infrastructure Oversee daily management, reliability, and performance of corporate IT systems and endpoints. Manage and expand a cloud-first IT infrastructure, including identity, device management, and collaboration tools. Direct the lifecycle management of employee hardware, software, and system access. AI Infrastructure and Emerging Technology Support and manage internal AI tools and infrastructure used throughout the company. Work with Engineering and Security teams to ensure responsible, secure, and scalable use of AI systems. Evaluate and implement AI-driven tools to improve productivity and automation for internal teams. SaaS and Identity Lifecycle Management Lead initiatives for SaaS and identity lifecycle management, focusing on integration and security.

Apr 29, 2026
Apply
TPC Engineering Holdings, LLC logoTPC Engineering Holdings, LLC logo
IT Infrastructure Specialist

TPC Engineering Holdings, LLC

Full-time|On-site|OH - Twinsburg | Trexon

Join our dynamic team as an IT Infrastructure Specialist, where you will play a crucial role in maintaining and enhancing our IT infrastructure. You will be responsible for ensuring the reliability, security, and performance of our IT systems. Your expertise will help us support our growing business operations effectively.

Apr 8, 2026
Apply
Percepta logoPercepta logo
Full-time|On-site|New York City

About UsAt Percepta, we are dedicated to revolutionizing vital sectors through the power of applied AI. Our goal is to ensure that key industries such as healthcare, manufacturing, and energy harness the benefits of cutting-edge technology.We partner with leading organizations to facilitate AI transformation, providing:Expertise in engineering, product development, and researchMosaic, our proprietary toolkit designed for the swift deployment of intelligent architecturesStrategic alliances with notable entities like Anthropic, McKinsey, AWS, and the General Catalyst portfolioOur team is a dynamic collective of Applied AI Engineers, Embedded Product Managers, and Researchers driven by the mission to integrate advanced AI into the systems that shape our world.Percepta is a proud partner of General Catalyst.Role OverviewWe are on the lookout for an AI Infrastructure Engineer who will be responsible for the infrastructure, deployment, and operational reliability that underpin Percepta's AI systems, including the autonomous agents driving our innovations.Your role will involve enhancing existing systems: refining our Terraform configurations, fortifying deployment pipelines, and implementing more robust management of infrastructure across various regions and providers. You will also be tasked with constructing missing components and exploring uncharted territories, defining what Site Reliability Engineering (SRE) means in the context of autonomous decision-making systems.The infrastructure paradigms for future autonomous systems are yet to be established, and you will play a crucial role in shaping them.What Sets This Role ApartYou will be working with autonomous systems, where the infrastructure dynamics shift significantly when workloads have agency.Observability entails understanding the rationale behind an agent's decisions, not merely checking the health of a pod.There is a tangible gap between research and production in our environment. Our teams transition optimization algorithms and AI systems from research settings to production, and you will be integral to this process. While MLOps experience is not mandatory, you will be closer to this boundary than most infrastructure roles.Join a small team with significant ownership. You will make foundational decisions rather than inherit pre-existing ones.Your ResponsibilitiesDesign infrastructure patterns for multi-agent systems that are observable, controllable, and recoverable in innovative ways.

Mar 30, 2026
Apply
Fireworks AI logoFireworks AI logo
Full-time|On-site|New York, NY; San Mateo, CA

About Us:At Fireworks AI, we are at the forefront of creating next-generation generative AI infrastructure. Our cutting-edge platform is recognized for delivering the highest-quality models with unparalleled speed and scalability in inference. Independently benchmarked as a leader in LLM inference speed, we drive significant advancements through innovative projects, including our proprietary function calling and multimodal models. As a Series C company valued at $4 billion and backed by leading investors such as Benchmark, Sequoia, Lightspeed, Index, and Evantic, we are a dynamic team of builders, comprised of veterans from Meta PyTorch and Google Vertex AI.The Role:We are seeking a talented Software Engineer to join our AI Infrastructure team. In this pivotal role, you will contribute to designing and developing the foundational systems that power Fireworks AI’s generative AI platform. Your focus will be on building robust infrastructure and tools that guarantee the reliability, performance, quality, and availability of our AI systems.Our mission is to establish Fireworks AI as the most dependable and user-friendly generative AI platform globally. You will collaborate closely with our cloud infrastructure, product, and performance teams to create infrastructure solutions that connect our customers with the high-performance proprietary Fireworks inference engine.Key Responsibilities:Design and develop scalable backend infrastructure supporting distributed training, inference, and data pipelines.Build and maintain essential backend services, including LLM CI/CD pipelines, control planes, and model serving systems.Enhance performance optimization, cost efficiency, and reliability across compute, storage, and networking layers.Create frameworks and safeguards to ensure Fireworks AI maintains the highest model quality in the industry.Work alongside performance, training, and product teams to translate research and product requirements into effective infrastructure solutions.Engage in code reviews, technical discussions, and continuous integration and deployment processes.

Mar 5, 2026
Apply
Spellbrush logoSpellbrush logo
Full-time|On-site|San Francisco

Join Our Team as an AI Infrastructure EngineerAt Spellbrush, the premier generative AI studio behind niji・journey, we are in search of a talented AI Infrastructure Engineer to help us develop and enhance our end-to-end machine learning infrastructure, facilitating the operation of our models across a variety of platforms.Key Responsibilities:Design, implement, and maintain next-generation inference architecture to optimize the performance of our models across mobile, web, and other platforms.Collaborate with a dynamic team focused on creating cutting-edge image generation models that serve over 16 million users globally.Ideal Candidate Profile:Experience with Large Distributed Systems: You possess a strong background in working with modern technologies such as Kubernetes (K8S), Kafka, NATS, Redis, among others. Your hands-on experience spans both on-premises and multi-cloud environments, and you understand the intricacies and potential pitfalls of each system.Expertise in GPU Workloads: Your understanding of GPU processing for handling substantial workloads sets you apart. Having experience in deploying or optimizing GPU workloads end-to-end is a significant advantage.Passion for Anime Aesthetics: As avid anime enthusiasts, we value team members who share our passion for the anime aesthetic, contributing to a creative movement that engages millions.Team Player in Fast-Paced Environments: You thrive in small, agile teams and are eager to work alongside some of the world's top AI researchers, contributing to the best image models globally. We believe in the power of in-person collaboration, with opportunities at our offices in Tokyo (downtown Akihabara) or San Francisco. Visa sponsorships are available.

Feb 7, 2024
Apply
Spear AI logo
Full-time|Hybrid|Washington, D.C. / Hybrid

Join Spear AI as a proficient Infrastructure Engineer, where you will contribute to the development of an advanced data management and AI platform aimed at enhancing maritime domain awareness.As an innovative defense contracting firm, Spear AI is committed to providing state-of-the-art solutions to bolster national security. We foster a culture that emphasizes innovation while fulfilling mission-critical objectives, operating within a flat organizational structure that empowers all employees to make a significant impact. Whether you join our Hardware, Software, or Services divisions, you will collaborate with highly skilled professionals dedicated to excellence and enhancing the capabilities that ensure our nation's safety and security.Our innovative sonobuoy sensors are deployed underwater to gather edge data, and we partner with the U.S. Navy to collect and process SONAR data. This role offers you the chance to work on impactful projects that directly influence warfighter capabilities and mission outcomes.

Apr 3, 2026
Apply
Eurofins Scientific logoEurofins Scientific logo
Full-time|On-site|Lancaster

Join Eurofins Scientific as an IT Operations and Infrastructure Specialist, where you will play a vital role in managing and enhancing our IT infrastructure. You will collaborate with cross-functional teams to ensure optimal performance and reliability of our IT systems.

Jul 14, 2020
Apply
Yutori logoYutori logo
Full-time|On-site|San Francisco, California, United States

At Yutori, we are transforming the way individuals engage with the digital realm by developing AI agents capable of efficiently performing everyday online tasks. Our approach is to create a comprehensive, agent-first ecosystem, encompassing everything from training proprietary models to designing innovative generative product interfaces.To further this mission, we are seeking a skilled AI Engineer to join our pioneering team. Ideal candidates should possess strong technical expertise and a passion for crafting superhuman AI agents that can navigate the web autonomously.Our founders — Devi Parikh, Abhishek Das, and Dhruv Batra — bring a wealth of experience in AI research and product development, particularly in generative, multimodal, and embodied AI, honed during their time at Meta. Our team merges AI proficiency with a design-oriented approach to advance Yutori’s objectives.Yutori is proudly supported by a distinguished group of visionary investors, including Elad Gil, Sarah Guo, Jeff Dean, Fei-Fei Li, Amjad Masad, Guillermo Rauch, Akshay Kothari, Soleio, Oliver Cameron, Julien Chaumond, Logan Kilpatrick, Bryan McCann, Vladlen Koltun, Jamie Cuffe, Michele Catasta, and many others.

Mar 26, 2025

Sign in to browse more jobs

Create account — see all 37,733 results

Tailoring 0 resumes

We'll move completed jobs to Ready to Apply automatically.