Clicking Apply Now takes you to AutoApply where you can tailor your resume and apply.
Unlock Your Potential
Generate Job-Optimized Resume
One Click And Our AI Optimizes Your Resume to Match The Job Description.
Is Your Resume Optimized For This Role?
Find Out If You're Highlighting The Right Skills And Fix What's Missing
Experience Level
Experience
Qualifications
What You’ll DoDesign, develop, and optimize machine learning models for our mobile applications. Research and apply cutting-edge AI techniques to enhance user engagement and app performance. Collaborate with cross-disciplinary teams to integrate AI solutions into our offerings. Establish and maintain scalable ML pipelines for efficient model deployment and monitoring. Analyze extensive datasets to extract insights and support data-driven strategies. Stay abreast of the latest AI trends and best practices, incorporating them into our development processes. Optimize AI models for mobile environments to ensure high performance and low latency.
About the job
Join Our Team at Air Apps
At Air Apps, we are on a mission to revolutionize resource management through innovative technology. Founded in 2018 in Lisbon, Portugal, we have expanded our reach with offices in both Lisbon and San Francisco, boasting over 100 million downloads globally. Our vision is to create the world’s first AI-powered Personal & Entrepreneurial Resource Planner (PRP), and we are looking for passionate individuals to help us achieve this goal.
Our commitment to challenging the status quo drives us to push the boundaries of AI-driven solutions that make a real impact. Here, you will have the opportunity to be a creative force, developing products that empower individuals worldwide.
Join us as we embark on this journey to redefine how people plan, work, and live.
About Air Apps
Air Apps is a forward-thinking technology company dedicated to creating innovative AI solutions that transform how people manage their resources. With a strong foundation built in Lisbon and an expanding presence in San Francisco, we are committed to delivering exceptional products that enhance productivity and efficiency for users globally.
Join Our Team at Air AppsAt Air Apps, we are on a mission to revolutionize resource management through innovative technology. Founded in 2018 in Lisbon, Portugal, we have expanded our reach with offices in both Lisbon and San Francisco, boasting over 100 million downloads globally. Our vision is to create the world’s first AI-powered Personal & Entrepreneurial Resource Planner (PRP), and we are looking for passionate individuals to help us achieve this goal.Our commitment to challenging the status quo drives us to push the boundaries of AI-driven solutions that make a real impact. Here, you will have the opportunity to be a creative force, developing products that empower individuals worldwide.Join us as we embark on this journey to redefine how people plan, work, and live.
Draup, based in San Francisco, is an AI company with Series A funding that builds intelligence solutions for large enterprises. The platform analyzes over 1 billion job descriptions and 850 million professional profiles, serving more than 250 enterprise clients, including several Fortune 10 companies. Draup’s data comes from more than 100 labor databases, supporting clients with deep workforce insights. Role overview The engineering team in Silicon Valley is expanding. Draup seeks experienced AI/ML Engineers interested in advancing both research and product development in artificial intelligence. What you will do Develop and maintain production-grade large language model (LLM) pipelines and agentic workflows. Design and enhance retrieval-augmented generation (RAG) architectures at scale, using vector databases such as Pinecone, FAISS, and Weaviate. Implement advanced agentic systems with tools like LangGraph or LlamaIndex, focusing on tool use, multi-agent coordination, and reasoning loops. Lead prompt engineering, manage model versioning, oversee evaluation (including RAGAS and DeepEval), and instrument LLMOps. Integrate AI features into large-scale data pipelines, ensuring observability in production and compliance with guardrails. Location This position is based in San Francisco, CA (Silicon Valley).
ML/AI Research Engineer - Founding Team at Agentic AI LabLocation: San Francisco Bay AreaType: Full-TimeCompensation: Competitive salary + meaningful equity (founding tier)At fabrion, backed by 8VC, we are assembling a top-tier team dedicated to addressing one of the most pressing infrastructure challenges in the industry.About the RoleJoin us in shaping the future of enterprise AI infrastructure, focusing on agents, retrieval-augmented generation (RAG), knowledge graphs, and multi-tenant governance.As an ML/AI Research Engineer, you will spearhead the design, training, evaluation, and optimization of agent-native AI models. Your work will integrate LLMs, vector search, graph reasoning, and reinforcement learning, establishing the intelligence layer for our enterprise data fabric.This role goes beyond prompt engineering; it encompasses the entire ML lifecycle—from data curation and fine-tuning to thorough evaluation, interpretability, and deployment, all while considering cost-effectiveness, alignment, and agent coordination.Core ResponsibilitiesFine-tune and assess open-source LLMs (e.g., LLaMA 3, Mistral, Falcon, Mixtral) for enterprise applications, leveraging both structured and unstructured data.Construct and enhance RAG pipelines utilizing LangChain, LangGraph, LlamaIndex, or Dust, integrating with our vector databases and internal knowledge graphs.Train agent architectures (ReAct, AutoGPT, BabyAGI, OpenAgents) using enterprise task datasets.Develop embedding-based memory and retrieval chains employing token-efficient chunking strategies.Create reinforcement learning pipelines to enhance agent behaviors (e.g., RLHF, DPO, PPO).Establish scalable evaluation harnesses for LLM and agent performance, including synthetic evaluations, trace capture, and explainability tools.Contribute to model observability, drift detection, error classification, and alignment efforts.Optimize inference latency and GPU resource utilization across both cloud and on-premises environments.Desired ExperienceModel Training:Deep understanding of machine learning principles and hands-on experience with model training.
Full-time|$308K/yr - $423.5K/yr|On-site|San Francisco, CA
About FaireFaire is a cutting-edge online wholesale marketplace driven by the belief that the future is local. Independent retailers around the world generate more revenue than giants like Walmart and Amazon combined, yet individually, they often struggle against these behemoths. At Faire, we harness the power of technology, data, and machine learning to connect this vibrant community of entrepreneurs globally. Imagine your favorite local boutique; we empower them to discover and sell exceptional products from around the world. With the right tools and insights, we aim to level the playing field, allowing small businesses to compete effectively with large retail chains and e-commerce platforms.By fostering the growth of independent businesses, Faire is making a positive economic impact in local communities worldwide. We’re in search of intelligent, resourceful, and passionate individuals to join us in driving the shop-local movement. If you share our belief in community, we would love to welcome you to ours.About this Role:We are on the lookout for a Principal ML / AI Engineer to serve as a company-wide technical thought leader and practitioner in shaping the future of Data and AI at Faire. This unique opportunity allows you to influence broad technical strategies across data, engineering, and product while engaging directly with pioneering AI research and applications. This role will report directly to the CTO of Faire.Your Responsibilities:Shape the AI Vision – Collaborate with product, design, strategy & analytics, machine learning, and the wider engineering leadership to define how AI can unlock transformational value for Faire’s retailers and brands. Provide thought leadership to guide company-wide priorities, particularly focusing on product strategy and key investment areas.Prototype and Unblock – Lead the development and implementation of AI systems (such as LLM fine-tuning, RLHF, agent frameworks, etc.) that illustrate what’s achievable and promote adoption across teams. Act as a “super individual contributor” who can delve deeply into technical challenges, enabling the engineering organization to advance quickly with AI and amplify both development and impact.Architect the AI-Ready Stack – Design Faire’s technical ecosystem, encompassing event logging, data warehouses, feature stores, and model serving, to ensure our infrastructure is AI-ready, scalable, and optimized for rapid experimentation.
Join our dynamic team at Cloudflare as a Senior/Principal Systems Engineer specializing in Workers AI (AI/ML). In this pivotal role, you will leverage your expertise in artificial intelligence and machine learning to develop cutting-edge solutions that enhance our platform's capabilities. You will collaborate with cross-functional teams to drive innovation and improve our systems, ensuring we remain at the forefront of technology.
Full-time|$208.6K/yr - $429.5K/yr|Remote|San Francisco, CA, US; Remote, US
About Pinterest:At Pinterest, our platform inspires millions of people around the globe to explore creative ideas, envision new possibilities, and create lasting memories. We are dedicated to providing the inspiration needed to build a fulfilling life, starting with the talented individuals who drive our product development.Join us in a career that sparks innovation for millions, transforms passion into opportunities for growth, and celebrates the diverse experiences of our team members, all while enjoying the flexibility to perform at your best. Building a career you love is within reach.Position Overview:We are looking for a Senior Engineering Manager to spearhead our AI/ML Serving Platform team, which develops the core tools and infrastructure utilized by numerous AI/ML engineers across Pinterest. This includes systems for recommendations, advertisements, visual search, notifications, and trust and safety. Our goal is to enhance the efficiency, quality, and speed of AI/ML systems, ensuring they are production-ready and reliable for iterative model development.Key Responsibilities:Lead the team in driving continuous improvements in advanced model architectures, optimizing resource usage, and boosting AI/ML developer productivity.Establish the technical vision for the team aligned with company and organizational priorities.Mentor and cultivate talent within the team.Qualifications:Proven experience in managing engineering teams with diverse cross-organizational clients.Expertise in developing large-scale distributed serving systems.Familiarity with AI/ML inference technologies (e.g., PyTorch, TensorFlow) for web-scale online serving.Bachelor's degree in Computer Science or a related field, or equivalent professional experience.
About Sygaldry Technologies Sygaldry Technologies develops quantum-accelerated AI servers in San Francisco, focusing on faster AI training and inference. By combining quantum technology with artificial intelligence, the team addresses challenges in computing costs and energy efficiency. Their AI servers integrate multiple qubit types within a fault-tolerant system, aiming for a balance of cost, scalability, and speed. The company values optimism, rigor, and a drive to solve complex problems in physics, engineering, and AI. Role Overview: ML Infrastructure Engineer The ML Infrastructure Engineer joins the AI & Algorithms team, which includes research scientists, applied mathematicians, and quantum algorithm specialists. This role centers on building and maintaining the compute infrastructure that powers advanced research. The systems you build will support reliable GPU access, reproducible experiments, and scalable workloads, so researchers can focus on their core work without needing deep cloud expertise. Expect to design and manage compute platforms for a range of tasks, including quantum circuit simulation, large-scale numerical optimization, model training, tensor network contractions, and high-throughput data generation. These workloads span multiple cloud providers and on-premises GPU servers. Key Responsibilities Develop compute abstractions for diverse workloads, such as GPU-accelerated simulations, distributed training, high-throughput CPU jobs, and interactive analyses using frameworks like PyTorch and JAX. Set up infrastructure to support experiment tracking and reproducibility. Create developer tools that make cloud computing feel local, streamlining environment setup, job submission, monitoring, and artifact management. Scale experiments from single-GPU prototypes to large, multi-node production runs. Multi-Cloud GPU Orchestration Design orchestration strategies for workloads across multiple cloud providers, optimizing job routing for cost, availability, and capability. Monitor and improve cloud spending, keeping track of credit balances, burn rates, and expiration dates.
Join Our Team as an Applied AI/ML Engineer!Are you an innovative AI/ML Engineer with 5 to 8 years of experience in developing and managing production AI and ML systems that impact real users? At paraform, we seek a talented individual who thrives in creating modern LLM agentic systems and traditional ML frameworks. You should have a proven track record of delivering systems with measurable quality metrics and a tangible business impact, alongside a strong grasp of evaluation methodologies that prioritize product goals. Familiarity with large-scale retrieval, ranking, NLP, or personalization systems is highly desirable.Your ResponsibilitiesDesign and implement cutting-edge AI/ML systems that enhance matchmaking, ranking, and automation within the Paraform marketplace.Oversee the complete model development lifecycle, from data collection and labeling strategies through to deployment, monitoring, and iterative improvements in production.Collaborate closely with product managers, AI engineers, and full-stack teams to deliver systems that significantly affect marketplace metrics.Establish and sustain robust evaluation frameworks to continuously assess quality, bias, reliability, and business impact.Set the technical direction and establish best practices for AI/ML at Paraform.Mentor fellow engineers, elevating the standards of design, deployment, and maintenance of AI systems.About paraform:Paraform is revolutionizing the hiring market—one of the largest and most fragmented industries globally. We collaborate with renowned industry leaders such as Cursor, Palantir, Windsurf, Decagon, Shopify, Coinbase, and Hightouch to attract world-class talent. Our rapid growth, exemplified by an 8x increase in revenue last year, underscores our success.As the first all-in-one AI recruiting marketplace, Paraform connects companies with thousands of specialized recruiters and AI agents to streamline the hiring process, making it faster and more precise.Our mission is to enhance hiring efficiency, reliability, and scalability for every organization worldwide.At paraform, we believe the best recruiting occurs when the right talent addresses the right challenges, blending human insight with AI that comprehends hiring intricacies. Hence, we are laying a new foundation for recruitment that unites specialized recruiters with advanced AI tools.
Full-time|$240K/yr - $270K/yr|On-site|San Francisco, CA
Role Overview Sigma Computing is building the next generation of data interaction. The platform lets users explore and analyze billions of data rows in seconds, all within a familiar spreadsheet-like interface. Sigma aims to make it simple to analyze, present, and build data-driven applications at scale. AI is central to Sigma's vision for the future. The company is expanding its use of artificial intelligence to help users build in Sigma, surface insights, and make decisions faster. What You Will Do As a Senior AI/ML Engineer, join a team focused on shaping the AI architecture behind Sigma's platform. This work directly impacts thousands of enterprises that depend on Sigma for their data workflows. The team is responsible for designing and implementing the systems that will power Sigma's AI-driven features for years to come. Location This position is based in San Francisco, CA.
At Runway ML, we are revolutionizing the intersection of art and science through innovative AI technology. Our mission is to build sophisticated world models that transcend traditional artificial intelligence limitations. We believe that to tackle the most pressing challenges—such as robotics, disease, and scientific breakthroughs—we need systems that can learn from experiences just like humans do. By simulating these experiences, we can expedite progress in ways that were previously unimaginable.Our diverse and driven team consists of creative thinkers who are passionate about pushing boundaries and achieving the extraordinary. If you share this ambition and are eager to contribute to our groundbreaking work, we invite you to join us.About the Role*We are open to hiring remotely across North America. We also have offices in NYC, San Francisco, and Seattle.We are on the lookout for a highly skilled and intellectually inquisitive Technical Accounting Manager to be our go-to authority on intricate accounting issues. This position offers significant visibility and is ideal for a professional adept at interpreting complex accounting guidelines, formulating sound conclusions, and translating technical insights into practical accounting practices.
ML Ops Engineer — Agentic AI Lab (Founding Team)Location: San Francisco Bay AreaType: Full-TimeCompensation: Competitive salary + meaningful equity (founding tier)At fabrion, supported by 8VC, we are assembling a premier team to address one of the industry's most significant infrastructure challenges.About the RoleJoin our AI Lab as we advance the future of intelligent infrastructure through pioneering open-source LLMs, agent-native pipelines, retrieval-augmented generation (RAG), and knowledge-graph-grounded models. We seek an ML Ops Engineer who will serve as the vital link between ML research and production systems, taking charge of automating the model training, deployment, versioning, and observability pipelines that empower our agents and AI data fabric.In this role, you will engage in compute orchestration, GPU infrastructure management, fine-tuned model lifecycle stewardship, and ensure model governance and security.Key ResponsibilitiesDesign and maintain secure, scalable, and automated pipelines for:LLM fine-tuning, SFT, LoRA, RLHF, and DPO trainingRAG embedding pipelines with real-time updatesModel conversion, quantization, and inference deploymentOversee hybrid compute infrastructure (cloud, on-premises, GPU clusters) for training and inference workloads utilizing Kubernetes, Ray, and TerraformContainerize models and agents using Docker, ensuring reproducible builds and CI/CD through GitHub Actions or ArgoCDEstablish and enforce model governance, including versioning, metadata management, lineage tracking, reproducibility, and evaluation captureDevelop and manage evaluation and benchmarking frameworks (e.g. OpenLLM-Evals, RAGAS, LangSmith)Integrate with security and access control mechanisms (OPA, ABAC, Keycloak) to implement model policies by tenantImplement observability practices for model latency, token usage, performance metrics, error tracing, and drift detectionAssist in deploying agentic applications using LangGraph, LangChain, and custom inference solutions.
Join Varomoney as a Principal AI/ML Architect, where you will lead groundbreaking projects that leverage artificial intelligence and machine learning to transform financial services. Your expertise will guide our engineering teams in developing innovative solutions that not only meet but exceed client expectations. You will be at the forefront of AI/ML technology, driving strategic initiatives and ensuring the highest standards of technical excellence.
At Sciforium, we are at the forefront of AI infrastructure, dedicated to the development of advanced multimodal AI models and an innovative serving platform that emphasizes high efficiency. With substantial funding and direct collaboration from AMD, our team is rapidly expanding to create the complete stack for pioneering AI models and dynamic real-time applications.Role OverviewThis position provides a distinct opportunity to engage with the fundamental systems that drive Sciforium's multimodal AI models. You will play a crucial role in constructing the model serving platform, working with C++, Python, runtime execution, and distributed infrastructure to design a swift, dependable engine for real-time AI applications.You will acquire practical experience in performance engineering, discover how large AI models are optimized and deployed at scale, and collaborate closely with ML researchers and seasoned systems engineers. If you thrive in low-level programming and are passionate about performance, this role offers both impactful contributions and significant growth opportunities.
Full-time|$218.4K/yr - $273K/yr|On-site|San Francisco, CA; Seattle, WA; New York, NY
Join Scale AI's ML platform team (RLXF) as a Machine Learning Research Engineer, where you will play a pivotal role in developing our advanced distributed framework for training and inference of large language models. This platform is vital for enabling machine learning engineers, researchers, data scientists, and operators to conduct rapid and automated training, as well as evaluation of LLMs and data quality.At Scale, we occupy a unique position in the AI landscape, serving as an essential provider of training and evaluation data along with comprehensive solutions for the entire ML lifecycle. You will collaborate closely with Scale's ML teams and researchers to enhance the foundational platform that underpins our ML research and development initiatives. Your contributions will be crucial in optimizing the platform to support the next generation of LLM training, inference, and data curation.If you are passionate about driving the future of AI through groundbreaking innovations, we want to hear from you!
About Liquid AILiquid AI, a pioneering company spun out of MIT CSAIL, is at the forefront of developing general-purpose AI systems that operate efficiently across various platforms, from data center accelerators to on-device hardware. Our commitment to low latency, minimal memory usage, privacy, and reliability allows us to partner with some of the most esteemed enterprises in consumer electronics, automotive, life sciences, and financial services. As we experience rapid growth, we are seeking exceptional talent to join our innovative journey.The OpportunityJoin our cutting-edge Audio team, where we are developing advanced speech-language models capable of handling Speech-to-Text (STT), Text-to-Speech (TTS), and speech-to-speech tasks within a unified architecture. This pivotal role supports applied audio model development, directly collaborating with the technical lead to deliver production systems that operate on-device under real-time constraints. You will take ownership of key workstreams encompassing data pipelines, evaluation systems, and customer deployments. If you are eager to tackle unique technical challenges within a small, elite team where your contributions are impactful, this is the role for you.What We're Looking ForWe are seeking an individual who:Builds first, theorizes later: You prioritize shipping working systems over theoretical models; production-grade code is your default.Owns outcomes end-to-end: You take full responsibility for everything from data pipelines to customer deployments and don't shy away from challenges.Thrives under constraints: On-device, low-latency, memory-constrained environments motivate you. You view constraints as opportunities for innovative design.Ramps quickly on new territory: You are comfortable closing knowledge gaps swiftly and actively seek feedback to drive results.The WorkDevelop and scale data pipelines for audio model training, including preprocessing, augmentation, and quality filtering at scale.Design, implement, and maintain evaluation systems that assess multimodal performance across both internal and public benchmarks.Fine-tune and adapt audio models to cater to customer-specific use cases, taking charge from requirement gathering through to deployment.Contribute production code to the core audio repository while collaborating closely with infrastructure and research teams.Facilitate experimentation under real hardware constraints, transitioning smoothly between customer-focused projects and core development initiatives.
About UsAt Lemurian Labs, we are dedicated to democratizing AI technology while prioritizing sustainability. Our mission is to create solutions that minimize environmental impact, ensuring that artificial intelligence serves humanity positively. We are committed to responsible innovation and the sustainable growth of AI.We are in the process of developing a state-of-the-art, portable compiler that empowers developers to 'build once, deploy anywhere.' This technology ensures seamless cross-platform integration, allowing for model training in the cloud and deployment at the edge, all while maximizing resource efficiency and scalability.If you are passionate about scaling AI sustainably and are eager to make AI development more powerful and accessible, we invite you to join our team at Lemurian Labs. Together, we can build a future that is innovative and responsible.The RoleWe are seeking a Senior ML Performance Engineer to take charge of designing and leading our Performance Testing Platform from inception. In this pivotal role, you will be recognized as the technical expert in measuring, validating, and enhancing the performance of large language models (including Llama 3.2 70B, DeepSeek, and others) prior to and following compiler optimization on cutting-edge GPU architectures.This is a critical position that will significantly impact our product quality and customer success. You will work at the intersection of Machine Learning systems, GPU architecture, and performance engineering, constructing the infrastructure that substantiates the value of our compiler.
About Our TeamAt OpenAI, our Hardware organization is pioneering the development of cutting-edge silicon and system-level solutions tailored to meet the distinctive needs of advanced AI workloads. We are dedicated to building the next generation of AI silicon, collaborating closely with software engineers and research partners to co-design hardware that integrates seamlessly with our AI models. Our mission includes not only delivering high-quality, production-grade silicon for OpenAI's supercomputing infrastructure but also creating custom design tools and methodologies that foster innovation and enable hardware optimized specifically for AI applications.About the RoleWe are on the lookout for a talented Research Hardware Co-Design Engineer to operate at the intersection of model research and silicon/system architecture. In this role, you will play a critical part in shaping the numerics, architecture, and technological strategies for the future of OpenAI's silicon in collaboration with both Research and Hardware teams.Your responsibilities will include diagnosing discrepancies between theoretical performance and real-world measurements, writing quantization kernels, assessing the risks associated with numerics through model evaluations, quantifying system architecture trade-offs, and implementing innovative numeric RTL. This is a hands-on position for individuals who are passionate about tackling challenging problems, seeking practical solutions, and driving them to production. Strong prioritization and transparent communication skills are vital for success in this role.Location: San Francisco, CA (Hybrid: 3 days/week onsite)Relocation assistance available.Key Responsibilities:Enhance our roofline simulator to monitor evolving workloads and deliver analyses that quantify the impact of architectural decisions, supporting technology exploration.Identify and resolve discrepancies between performance simulations and actual measurements; effectively communicate root causes, bottlenecks, and incorrect assumptions.Develop emulation kernels for low-precision numerics and lossy compression techniques, equipping Research with the insights needed to balance efficiency with model quality.Prototype numeric modules by advancing RTL through synthesis; either hand off innovative numeric solutions cleanly or occasionally take ownership of an RTL module from start to finish.Proactively engage with new ML workloads, prototype them using rooflines and/or functional simulations, and initiate evaluations of new opportunities or risks.Gain a holistic understanding of the transition from ML science to hardware optimization, breaking down this comprehensive objective into actionable short-term deliverables.Foster collaborative relationships across diverse teams with varying goals and expertise, ensuring that progress remains unimpeded.Clearly articulate design trade-offs with explicit assumptions and rationale.
Join AfterQuery as an AI/ML Research Intern and immerse yourself in groundbreaking artificial intelligence projects. This internship is designed for exceptional undergraduate and master's students eager to collaborate with our research team on advanced reasoning and agentic models. You will have the opportunity to access specialized datasets and work closely with industry experts, contributing to exciting AI research that could lead to co-authored papers and presentations at prestigious AI conferences.We invite students currently enrolled in relevant programs to apply. This role requires a commitment of 10 to 40 hours per week, adaptable to the needs of the company.
Full-time|$225K/yr - $315K/yr|Remote|San Francisco
About the CompanyLavendo is a pioneering publicly traded company leading the charge in the AI revolution. With an AI-centric cloud platform, we are transforming the artificial intelligence landscape. Our state-of-the-art infrastructure, including extensive GPU clusters and advanced cloud services, supports developers in harnessing the explosive growth of the global AI industry, catering to Fortune 1000 firms, innovative startups, and AI researchers alike.Company type: Publicly tradedIndustry: AI/ML, Cloud Computing, Infrastructure-as-CodeCandidate Location: Remote U.S.Our mission is to democratize AI infrastructure access and empower organizations to innovate, optimize, and deploy AI solutions seamlessly at any scale. By simplifying the complexities of AI development, we provide a comprehensive full-stack AI platform that marries robust hardware with easy-to-use tools and services.The OpportunityWe are on the lookout for a Senior AI/ML Specialist Solutions Architect to become a crucial part of our client's dynamic team. This role presents an exciting opportunity to design and implement scalable AI solutions tailored for AI-centric clients, leveraging cutting-edge technologies and contributing to one of the most powerful commercially available supercomputers.What You'll DoArchitect and enhance distributed training and inference systems for large-scale AI models.Design and deliver customer-centric solutions that optimize performance and drive business value.Lead the migration of ML pipelines from Proof of Concept to scalable production environments.Foster long-term relationships with clients, ensuring satisfaction and alignment with their strategic objectives.Produce whitepapers, conduct technical presentations, and facilitate webinars to disseminate insights and best practices.Provide technical guidance and mentorship to teams regarding AI infrastructure and deployment strategies.Collaborate with engineering and product teams to prioritize customer feedback and shape product roadmaps.
About AbridgeAbridge, founded in 2018, is dedicated to enhancing understanding within healthcare. Our innovative AI-driven platform is designed specifically for medical conversations, streamlining clinical documentation while allowing clinicians to concentrate on their primary focus—their patients.Our robust technology converts patient-clinician dialogues into structured clinical notes in real-time, integrating seamlessly with EMR systems. Thanks to Linked Evidence and our specially designed, auditable AI, we uniquely align AI-generated summaries with ground truth, enabling providers to trust and verify outputs swiftly. As trailblazers in generative AI for healthcare, we are establishing industry standards for the ethical application of AI across health systems.Our diverse team comprises practicing MDs, AI scientists, PhDs, creatives, technologists, and engineers collaborating to empower individuals and enhance the clarity of care. Our offices are located in the vibrant Mission District of San Francisco, the bustling SoHo area of New York, and East Liberty in Pittsburgh.The RoleThe success of Abridge hinges on delivering the most accurate, personalized, and clinically relevant notes, forming the cornerstone of our product strategy. The quality of our notes is vital for clinician trust and our ability to expand into additional workflows. This role is pivotal in ensuring the quality and personalization of all notes.The Note Generation team is central to this mission, overseeing the models that drive documentation, quality measurement systems at scale, and specialty and workflow engines that guarantee our outputs align with diverse clinical practices.We seek a visionary product leader to direct the product vision, strategy, and execution for Note Generation. You will lead a team of Product Managers and collaborate closely with engineering, machine learning, design, clinical leaders, and our specialty councils. You will establish the standard for note quality, shape our model and evaluation roadmap, and ensure we deliver both essential and distinctive capabilities required to dominate the market.What You’ll DoLead AI product strategy for Note Generation. Define the long-term strategy for Abridge’s core note models and evaluation stack.Oversee the evaluation platform and measurement systems. Create a world-class note quality evaluation system utilizing human annotation, LLM judges, automated backtesting, and production analytics. Collaborate with ML and engineering teams to ensure reliable benchmarks, continuous validation, and rapid iteration cycles.Drive innovation and best practices. Continuously enhance our product offerings and lead cross-functional initiatives to ensure our solutions remain at the forefront of AI in healthcare.
Join Our Team at Air AppsAt Air Apps, we are on a mission to revolutionize resource management through innovative technology. Founded in 2018 in Lisbon, Portugal, we have expanded our reach with offices in both Lisbon and San Francisco, boasting over 100 million downloads globally. Our vision is to create the world’s first AI-powered Personal & Entrepreneurial Resource Planner (PRP), and we are looking for passionate individuals to help us achieve this goal.Our commitment to challenging the status quo drives us to push the boundaries of AI-driven solutions that make a real impact. Here, you will have the opportunity to be a creative force, developing products that empower individuals worldwide.Join us as we embark on this journey to redefine how people plan, work, and live.
Draup, based in San Francisco, is an AI company with Series A funding that builds intelligence solutions for large enterprises. The platform analyzes over 1 billion job descriptions and 850 million professional profiles, serving more than 250 enterprise clients, including several Fortune 10 companies. Draup’s data comes from more than 100 labor databases, supporting clients with deep workforce insights. Role overview The engineering team in Silicon Valley is expanding. Draup seeks experienced AI/ML Engineers interested in advancing both research and product development in artificial intelligence. What you will do Develop and maintain production-grade large language model (LLM) pipelines and agentic workflows. Design and enhance retrieval-augmented generation (RAG) architectures at scale, using vector databases such as Pinecone, FAISS, and Weaviate. Implement advanced agentic systems with tools like LangGraph or LlamaIndex, focusing on tool use, multi-agent coordination, and reasoning loops. Lead prompt engineering, manage model versioning, oversee evaluation (including RAGAS and DeepEval), and instrument LLMOps. Integrate AI features into large-scale data pipelines, ensuring observability in production and compliance with guardrails. Location This position is based in San Francisco, CA (Silicon Valley).
ML/AI Research Engineer - Founding Team at Agentic AI LabLocation: San Francisco Bay AreaType: Full-TimeCompensation: Competitive salary + meaningful equity (founding tier)At fabrion, backed by 8VC, we are assembling a top-tier team dedicated to addressing one of the most pressing infrastructure challenges in the industry.About the RoleJoin us in shaping the future of enterprise AI infrastructure, focusing on agents, retrieval-augmented generation (RAG), knowledge graphs, and multi-tenant governance.As an ML/AI Research Engineer, you will spearhead the design, training, evaluation, and optimization of agent-native AI models. Your work will integrate LLMs, vector search, graph reasoning, and reinforcement learning, establishing the intelligence layer for our enterprise data fabric.This role goes beyond prompt engineering; it encompasses the entire ML lifecycle—from data curation and fine-tuning to thorough evaluation, interpretability, and deployment, all while considering cost-effectiveness, alignment, and agent coordination.Core ResponsibilitiesFine-tune and assess open-source LLMs (e.g., LLaMA 3, Mistral, Falcon, Mixtral) for enterprise applications, leveraging both structured and unstructured data.Construct and enhance RAG pipelines utilizing LangChain, LangGraph, LlamaIndex, or Dust, integrating with our vector databases and internal knowledge graphs.Train agent architectures (ReAct, AutoGPT, BabyAGI, OpenAgents) using enterprise task datasets.Develop embedding-based memory and retrieval chains employing token-efficient chunking strategies.Create reinforcement learning pipelines to enhance agent behaviors (e.g., RLHF, DPO, PPO).Establish scalable evaluation harnesses for LLM and agent performance, including synthetic evaluations, trace capture, and explainability tools.Contribute to model observability, drift detection, error classification, and alignment efforts.Optimize inference latency and GPU resource utilization across both cloud and on-premises environments.Desired ExperienceModel Training:Deep understanding of machine learning principles and hands-on experience with model training.
Full-time|$308K/yr - $423.5K/yr|On-site|San Francisco, CA
About FaireFaire is a cutting-edge online wholesale marketplace driven by the belief that the future is local. Independent retailers around the world generate more revenue than giants like Walmart and Amazon combined, yet individually, they often struggle against these behemoths. At Faire, we harness the power of technology, data, and machine learning to connect this vibrant community of entrepreneurs globally. Imagine your favorite local boutique; we empower them to discover and sell exceptional products from around the world. With the right tools and insights, we aim to level the playing field, allowing small businesses to compete effectively with large retail chains and e-commerce platforms.By fostering the growth of independent businesses, Faire is making a positive economic impact in local communities worldwide. We’re in search of intelligent, resourceful, and passionate individuals to join us in driving the shop-local movement. If you share our belief in community, we would love to welcome you to ours.About this Role:We are on the lookout for a Principal ML / AI Engineer to serve as a company-wide technical thought leader and practitioner in shaping the future of Data and AI at Faire. This unique opportunity allows you to influence broad technical strategies across data, engineering, and product while engaging directly with pioneering AI research and applications. This role will report directly to the CTO of Faire.Your Responsibilities:Shape the AI Vision – Collaborate with product, design, strategy & analytics, machine learning, and the wider engineering leadership to define how AI can unlock transformational value for Faire’s retailers and brands. Provide thought leadership to guide company-wide priorities, particularly focusing on product strategy and key investment areas.Prototype and Unblock – Lead the development and implementation of AI systems (such as LLM fine-tuning, RLHF, agent frameworks, etc.) that illustrate what’s achievable and promote adoption across teams. Act as a “super individual contributor” who can delve deeply into technical challenges, enabling the engineering organization to advance quickly with AI and amplify both development and impact.Architect the AI-Ready Stack – Design Faire’s technical ecosystem, encompassing event logging, data warehouses, feature stores, and model serving, to ensure our infrastructure is AI-ready, scalable, and optimized for rapid experimentation.
Join our dynamic team at Cloudflare as a Senior/Principal Systems Engineer specializing in Workers AI (AI/ML). In this pivotal role, you will leverage your expertise in artificial intelligence and machine learning to develop cutting-edge solutions that enhance our platform's capabilities. You will collaborate with cross-functional teams to drive innovation and improve our systems, ensuring we remain at the forefront of technology.
Full-time|$208.6K/yr - $429.5K/yr|Remote|San Francisco, CA, US; Remote, US
About Pinterest:At Pinterest, our platform inspires millions of people around the globe to explore creative ideas, envision new possibilities, and create lasting memories. We are dedicated to providing the inspiration needed to build a fulfilling life, starting with the talented individuals who drive our product development.Join us in a career that sparks innovation for millions, transforms passion into opportunities for growth, and celebrates the diverse experiences of our team members, all while enjoying the flexibility to perform at your best. Building a career you love is within reach.Position Overview:We are looking for a Senior Engineering Manager to spearhead our AI/ML Serving Platform team, which develops the core tools and infrastructure utilized by numerous AI/ML engineers across Pinterest. This includes systems for recommendations, advertisements, visual search, notifications, and trust and safety. Our goal is to enhance the efficiency, quality, and speed of AI/ML systems, ensuring they are production-ready and reliable for iterative model development.Key Responsibilities:Lead the team in driving continuous improvements in advanced model architectures, optimizing resource usage, and boosting AI/ML developer productivity.Establish the technical vision for the team aligned with company and organizational priorities.Mentor and cultivate talent within the team.Qualifications:Proven experience in managing engineering teams with diverse cross-organizational clients.Expertise in developing large-scale distributed serving systems.Familiarity with AI/ML inference technologies (e.g., PyTorch, TensorFlow) for web-scale online serving.Bachelor's degree in Computer Science or a related field, or equivalent professional experience.
About Sygaldry Technologies Sygaldry Technologies develops quantum-accelerated AI servers in San Francisco, focusing on faster AI training and inference. By combining quantum technology with artificial intelligence, the team addresses challenges in computing costs and energy efficiency. Their AI servers integrate multiple qubit types within a fault-tolerant system, aiming for a balance of cost, scalability, and speed. The company values optimism, rigor, and a drive to solve complex problems in physics, engineering, and AI. Role Overview: ML Infrastructure Engineer The ML Infrastructure Engineer joins the AI & Algorithms team, which includes research scientists, applied mathematicians, and quantum algorithm specialists. This role centers on building and maintaining the compute infrastructure that powers advanced research. The systems you build will support reliable GPU access, reproducible experiments, and scalable workloads, so researchers can focus on their core work without needing deep cloud expertise. Expect to design and manage compute platforms for a range of tasks, including quantum circuit simulation, large-scale numerical optimization, model training, tensor network contractions, and high-throughput data generation. These workloads span multiple cloud providers and on-premises GPU servers. Key Responsibilities Develop compute abstractions for diverse workloads, such as GPU-accelerated simulations, distributed training, high-throughput CPU jobs, and interactive analyses using frameworks like PyTorch and JAX. Set up infrastructure to support experiment tracking and reproducibility. Create developer tools that make cloud computing feel local, streamlining environment setup, job submission, monitoring, and artifact management. Scale experiments from single-GPU prototypes to large, multi-node production runs. Multi-Cloud GPU Orchestration Design orchestration strategies for workloads across multiple cloud providers, optimizing job routing for cost, availability, and capability. Monitor and improve cloud spending, keeping track of credit balances, burn rates, and expiration dates.
Join Our Team as an Applied AI/ML Engineer!Are you an innovative AI/ML Engineer with 5 to 8 years of experience in developing and managing production AI and ML systems that impact real users? At paraform, we seek a talented individual who thrives in creating modern LLM agentic systems and traditional ML frameworks. You should have a proven track record of delivering systems with measurable quality metrics and a tangible business impact, alongside a strong grasp of evaluation methodologies that prioritize product goals. Familiarity with large-scale retrieval, ranking, NLP, or personalization systems is highly desirable.Your ResponsibilitiesDesign and implement cutting-edge AI/ML systems that enhance matchmaking, ranking, and automation within the Paraform marketplace.Oversee the complete model development lifecycle, from data collection and labeling strategies through to deployment, monitoring, and iterative improvements in production.Collaborate closely with product managers, AI engineers, and full-stack teams to deliver systems that significantly affect marketplace metrics.Establish and sustain robust evaluation frameworks to continuously assess quality, bias, reliability, and business impact.Set the technical direction and establish best practices for AI/ML at Paraform.Mentor fellow engineers, elevating the standards of design, deployment, and maintenance of AI systems.About paraform:Paraform is revolutionizing the hiring market—one of the largest and most fragmented industries globally. We collaborate with renowned industry leaders such as Cursor, Palantir, Windsurf, Decagon, Shopify, Coinbase, and Hightouch to attract world-class talent. Our rapid growth, exemplified by an 8x increase in revenue last year, underscores our success.As the first all-in-one AI recruiting marketplace, Paraform connects companies with thousands of specialized recruiters and AI agents to streamline the hiring process, making it faster and more precise.Our mission is to enhance hiring efficiency, reliability, and scalability for every organization worldwide.At paraform, we believe the best recruiting occurs when the right talent addresses the right challenges, blending human insight with AI that comprehends hiring intricacies. Hence, we are laying a new foundation for recruitment that unites specialized recruiters with advanced AI tools.
Full-time|$240K/yr - $270K/yr|On-site|San Francisco, CA
Role Overview Sigma Computing is building the next generation of data interaction. The platform lets users explore and analyze billions of data rows in seconds, all within a familiar spreadsheet-like interface. Sigma aims to make it simple to analyze, present, and build data-driven applications at scale. AI is central to Sigma's vision for the future. The company is expanding its use of artificial intelligence to help users build in Sigma, surface insights, and make decisions faster. What You Will Do As a Senior AI/ML Engineer, join a team focused on shaping the AI architecture behind Sigma's platform. This work directly impacts thousands of enterprises that depend on Sigma for their data workflows. The team is responsible for designing and implementing the systems that will power Sigma's AI-driven features for years to come. Location This position is based in San Francisco, CA.
At Runway ML, we are revolutionizing the intersection of art and science through innovative AI technology. Our mission is to build sophisticated world models that transcend traditional artificial intelligence limitations. We believe that to tackle the most pressing challenges—such as robotics, disease, and scientific breakthroughs—we need systems that can learn from experiences just like humans do. By simulating these experiences, we can expedite progress in ways that were previously unimaginable.Our diverse and driven team consists of creative thinkers who are passionate about pushing boundaries and achieving the extraordinary. If you share this ambition and are eager to contribute to our groundbreaking work, we invite you to join us.About the Role*We are open to hiring remotely across North America. We also have offices in NYC, San Francisco, and Seattle.We are on the lookout for a highly skilled and intellectually inquisitive Technical Accounting Manager to be our go-to authority on intricate accounting issues. This position offers significant visibility and is ideal for a professional adept at interpreting complex accounting guidelines, formulating sound conclusions, and translating technical insights into practical accounting practices.
ML Ops Engineer — Agentic AI Lab (Founding Team)Location: San Francisco Bay AreaType: Full-TimeCompensation: Competitive salary + meaningful equity (founding tier)At fabrion, supported by 8VC, we are assembling a premier team to address one of the industry's most significant infrastructure challenges.About the RoleJoin our AI Lab as we advance the future of intelligent infrastructure through pioneering open-source LLMs, agent-native pipelines, retrieval-augmented generation (RAG), and knowledge-graph-grounded models. We seek an ML Ops Engineer who will serve as the vital link between ML research and production systems, taking charge of automating the model training, deployment, versioning, and observability pipelines that empower our agents and AI data fabric.In this role, you will engage in compute orchestration, GPU infrastructure management, fine-tuned model lifecycle stewardship, and ensure model governance and security.Key ResponsibilitiesDesign and maintain secure, scalable, and automated pipelines for:LLM fine-tuning, SFT, LoRA, RLHF, and DPO trainingRAG embedding pipelines with real-time updatesModel conversion, quantization, and inference deploymentOversee hybrid compute infrastructure (cloud, on-premises, GPU clusters) for training and inference workloads utilizing Kubernetes, Ray, and TerraformContainerize models and agents using Docker, ensuring reproducible builds and CI/CD through GitHub Actions or ArgoCDEstablish and enforce model governance, including versioning, metadata management, lineage tracking, reproducibility, and evaluation captureDevelop and manage evaluation and benchmarking frameworks (e.g. OpenLLM-Evals, RAGAS, LangSmith)Integrate with security and access control mechanisms (OPA, ABAC, Keycloak) to implement model policies by tenantImplement observability practices for model latency, token usage, performance metrics, error tracing, and drift detectionAssist in deploying agentic applications using LangGraph, LangChain, and custom inference solutions.
Join Varomoney as a Principal AI/ML Architect, where you will lead groundbreaking projects that leverage artificial intelligence and machine learning to transform financial services. Your expertise will guide our engineering teams in developing innovative solutions that not only meet but exceed client expectations. You will be at the forefront of AI/ML technology, driving strategic initiatives and ensuring the highest standards of technical excellence.
At Sciforium, we are at the forefront of AI infrastructure, dedicated to the development of advanced multimodal AI models and an innovative serving platform that emphasizes high efficiency. With substantial funding and direct collaboration from AMD, our team is rapidly expanding to create the complete stack for pioneering AI models and dynamic real-time applications.Role OverviewThis position provides a distinct opportunity to engage with the fundamental systems that drive Sciforium's multimodal AI models. You will play a crucial role in constructing the model serving platform, working with C++, Python, runtime execution, and distributed infrastructure to design a swift, dependable engine for real-time AI applications.You will acquire practical experience in performance engineering, discover how large AI models are optimized and deployed at scale, and collaborate closely with ML researchers and seasoned systems engineers. If you thrive in low-level programming and are passionate about performance, this role offers both impactful contributions and significant growth opportunities.
Full-time|$218.4K/yr - $273K/yr|On-site|San Francisco, CA; Seattle, WA; New York, NY
Join Scale AI's ML platform team (RLXF) as a Machine Learning Research Engineer, where you will play a pivotal role in developing our advanced distributed framework for training and inference of large language models. This platform is vital for enabling machine learning engineers, researchers, data scientists, and operators to conduct rapid and automated training, as well as evaluation of LLMs and data quality.At Scale, we occupy a unique position in the AI landscape, serving as an essential provider of training and evaluation data along with comprehensive solutions for the entire ML lifecycle. You will collaborate closely with Scale's ML teams and researchers to enhance the foundational platform that underpins our ML research and development initiatives. Your contributions will be crucial in optimizing the platform to support the next generation of LLM training, inference, and data curation.If you are passionate about driving the future of AI through groundbreaking innovations, we want to hear from you!
About Liquid AILiquid AI, a pioneering company spun out of MIT CSAIL, is at the forefront of developing general-purpose AI systems that operate efficiently across various platforms, from data center accelerators to on-device hardware. Our commitment to low latency, minimal memory usage, privacy, and reliability allows us to partner with some of the most esteemed enterprises in consumer electronics, automotive, life sciences, and financial services. As we experience rapid growth, we are seeking exceptional talent to join our innovative journey.The OpportunityJoin our cutting-edge Audio team, where we are developing advanced speech-language models capable of handling Speech-to-Text (STT), Text-to-Speech (TTS), and speech-to-speech tasks within a unified architecture. This pivotal role supports applied audio model development, directly collaborating with the technical lead to deliver production systems that operate on-device under real-time constraints. You will take ownership of key workstreams encompassing data pipelines, evaluation systems, and customer deployments. If you are eager to tackle unique technical challenges within a small, elite team where your contributions are impactful, this is the role for you.What We're Looking ForWe are seeking an individual who:Builds first, theorizes later: You prioritize shipping working systems over theoretical models; production-grade code is your default.Owns outcomes end-to-end: You take full responsibility for everything from data pipelines to customer deployments and don't shy away from challenges.Thrives under constraints: On-device, low-latency, memory-constrained environments motivate you. You view constraints as opportunities for innovative design.Ramps quickly on new territory: You are comfortable closing knowledge gaps swiftly and actively seek feedback to drive results.The WorkDevelop and scale data pipelines for audio model training, including preprocessing, augmentation, and quality filtering at scale.Design, implement, and maintain evaluation systems that assess multimodal performance across both internal and public benchmarks.Fine-tune and adapt audio models to cater to customer-specific use cases, taking charge from requirement gathering through to deployment.Contribute production code to the core audio repository while collaborating closely with infrastructure and research teams.Facilitate experimentation under real hardware constraints, transitioning smoothly between customer-focused projects and core development initiatives.
About UsAt Lemurian Labs, we are dedicated to democratizing AI technology while prioritizing sustainability. Our mission is to create solutions that minimize environmental impact, ensuring that artificial intelligence serves humanity positively. We are committed to responsible innovation and the sustainable growth of AI.We are in the process of developing a state-of-the-art, portable compiler that empowers developers to 'build once, deploy anywhere.' This technology ensures seamless cross-platform integration, allowing for model training in the cloud and deployment at the edge, all while maximizing resource efficiency and scalability.If you are passionate about scaling AI sustainably and are eager to make AI development more powerful and accessible, we invite you to join our team at Lemurian Labs. Together, we can build a future that is innovative and responsible.The RoleWe are seeking a Senior ML Performance Engineer to take charge of designing and leading our Performance Testing Platform from inception. In this pivotal role, you will be recognized as the technical expert in measuring, validating, and enhancing the performance of large language models (including Llama 3.2 70B, DeepSeek, and others) prior to and following compiler optimization on cutting-edge GPU architectures.This is a critical position that will significantly impact our product quality and customer success. You will work at the intersection of Machine Learning systems, GPU architecture, and performance engineering, constructing the infrastructure that substantiates the value of our compiler.
About Our TeamAt OpenAI, our Hardware organization is pioneering the development of cutting-edge silicon and system-level solutions tailored to meet the distinctive needs of advanced AI workloads. We are dedicated to building the next generation of AI silicon, collaborating closely with software engineers and research partners to co-design hardware that integrates seamlessly with our AI models. Our mission includes not only delivering high-quality, production-grade silicon for OpenAI's supercomputing infrastructure but also creating custom design tools and methodologies that foster innovation and enable hardware optimized specifically for AI applications.About the RoleWe are on the lookout for a talented Research Hardware Co-Design Engineer to operate at the intersection of model research and silicon/system architecture. In this role, you will play a critical part in shaping the numerics, architecture, and technological strategies for the future of OpenAI's silicon in collaboration with both Research and Hardware teams.Your responsibilities will include diagnosing discrepancies between theoretical performance and real-world measurements, writing quantization kernels, assessing the risks associated with numerics through model evaluations, quantifying system architecture trade-offs, and implementing innovative numeric RTL. This is a hands-on position for individuals who are passionate about tackling challenging problems, seeking practical solutions, and driving them to production. Strong prioritization and transparent communication skills are vital for success in this role.Location: San Francisco, CA (Hybrid: 3 days/week onsite)Relocation assistance available.Key Responsibilities:Enhance our roofline simulator to monitor evolving workloads and deliver analyses that quantify the impact of architectural decisions, supporting technology exploration.Identify and resolve discrepancies between performance simulations and actual measurements; effectively communicate root causes, bottlenecks, and incorrect assumptions.Develop emulation kernels for low-precision numerics and lossy compression techniques, equipping Research with the insights needed to balance efficiency with model quality.Prototype numeric modules by advancing RTL through synthesis; either hand off innovative numeric solutions cleanly or occasionally take ownership of an RTL module from start to finish.Proactively engage with new ML workloads, prototype them using rooflines and/or functional simulations, and initiate evaluations of new opportunities or risks.Gain a holistic understanding of the transition from ML science to hardware optimization, breaking down this comprehensive objective into actionable short-term deliverables.Foster collaborative relationships across diverse teams with varying goals and expertise, ensuring that progress remains unimpeded.Clearly articulate design trade-offs with explicit assumptions and rationale.
Join AfterQuery as an AI/ML Research Intern and immerse yourself in groundbreaking artificial intelligence projects. This internship is designed for exceptional undergraduate and master's students eager to collaborate with our research team on advanced reasoning and agentic models. You will have the opportunity to access specialized datasets and work closely with industry experts, contributing to exciting AI research that could lead to co-authored papers and presentations at prestigious AI conferences.We invite students currently enrolled in relevant programs to apply. This role requires a commitment of 10 to 40 hours per week, adaptable to the needs of the company.
Full-time|$225K/yr - $315K/yr|Remote|San Francisco
About the CompanyLavendo is a pioneering publicly traded company leading the charge in the AI revolution. With an AI-centric cloud platform, we are transforming the artificial intelligence landscape. Our state-of-the-art infrastructure, including extensive GPU clusters and advanced cloud services, supports developers in harnessing the explosive growth of the global AI industry, catering to Fortune 1000 firms, innovative startups, and AI researchers alike.Company type: Publicly tradedIndustry: AI/ML, Cloud Computing, Infrastructure-as-CodeCandidate Location: Remote U.S.Our mission is to democratize AI infrastructure access and empower organizations to innovate, optimize, and deploy AI solutions seamlessly at any scale. By simplifying the complexities of AI development, we provide a comprehensive full-stack AI platform that marries robust hardware with easy-to-use tools and services.The OpportunityWe are on the lookout for a Senior AI/ML Specialist Solutions Architect to become a crucial part of our client's dynamic team. This role presents an exciting opportunity to design and implement scalable AI solutions tailored for AI-centric clients, leveraging cutting-edge technologies and contributing to one of the most powerful commercially available supercomputers.What You'll DoArchitect and enhance distributed training and inference systems for large-scale AI models.Design and deliver customer-centric solutions that optimize performance and drive business value.Lead the migration of ML pipelines from Proof of Concept to scalable production environments.Foster long-term relationships with clients, ensuring satisfaction and alignment with their strategic objectives.Produce whitepapers, conduct technical presentations, and facilitate webinars to disseminate insights and best practices.Provide technical guidance and mentorship to teams regarding AI infrastructure and deployment strategies.Collaborate with engineering and product teams to prioritize customer feedback and shape product roadmaps.
About AbridgeAbridge, founded in 2018, is dedicated to enhancing understanding within healthcare. Our innovative AI-driven platform is designed specifically for medical conversations, streamlining clinical documentation while allowing clinicians to concentrate on their primary focus—their patients.Our robust technology converts patient-clinician dialogues into structured clinical notes in real-time, integrating seamlessly with EMR systems. Thanks to Linked Evidence and our specially designed, auditable AI, we uniquely align AI-generated summaries with ground truth, enabling providers to trust and verify outputs swiftly. As trailblazers in generative AI for healthcare, we are establishing industry standards for the ethical application of AI across health systems.Our diverse team comprises practicing MDs, AI scientists, PhDs, creatives, technologists, and engineers collaborating to empower individuals and enhance the clarity of care. Our offices are located in the vibrant Mission District of San Francisco, the bustling SoHo area of New York, and East Liberty in Pittsburgh.The RoleThe success of Abridge hinges on delivering the most accurate, personalized, and clinically relevant notes, forming the cornerstone of our product strategy. The quality of our notes is vital for clinician trust and our ability to expand into additional workflows. This role is pivotal in ensuring the quality and personalization of all notes.The Note Generation team is central to this mission, overseeing the models that drive documentation, quality measurement systems at scale, and specialty and workflow engines that guarantee our outputs align with diverse clinical practices.We seek a visionary product leader to direct the product vision, strategy, and execution for Note Generation. You will lead a team of Product Managers and collaborate closely with engineering, machine learning, design, clinical leaders, and our specialty councils. You will establish the standard for note quality, shape our model and evaluation roadmap, and ensure we deliver both essential and distinctive capabilities required to dominate the market.What You’ll DoLead AI product strategy for Note Generation. Define the long-term strategy for Abridge’s core note models and evaluation stack.Oversee the evaluation platform and measurement systems. Create a world-class note quality evaluation system utilizing human annotation, LLM judges, automated backtesting, and production analytics. Collaborate with ML and engineering teams to ensure reliable benchmarks, continuous validation, and rapid iteration cycles.Drive innovation and best practices. Continuously enhance our product offerings and lead cross-functional initiatives to ensure our solutions remain at the forefront of AI in healthcare.
Dec 10, 2025
Sign in to browse more jobs
Create account — see all 7,271 results
Tailoring 0 resumes…
Tailoring 0 resumes…
We'll move completed jobs to Ready to Apply automatically.