Clicking Apply Now takes you to AutoApply where you can tailor your resume and apply.
Unlock Your Potential
Generate Job-Optimized Resume
One Click And Our AI Optimizes Your Resume to Match The Job Description.
Is Your Resume Optimized For This Role?
Find Out If You're Highlighting The Right Skills And Fix What's Missing
Experience Level
Senior
Qualifications
Proven experience in software engineering with a focus on infrastructure. Strong understanding of cloud services, networking, and system architecture. Proficiency in programming languages such as Python, Go, or Java. Experience with containerization technologies like Docker and Kubernetes. Excellent problem-solving skills and the ability to work in a fast-paced environment.
About the job
Join our dynamic team at Parafin as a Senior Software Engineer specializing in Infrastructure. In this pivotal role, you will design, develop, and maintain robust infrastructure solutions that support our scalable applications. Your expertise will help us enhance system performance, reliability, and security.
We are looking for innovative thinkers who thrive in a collaborative environment. You will work closely with cross-functional teams to implement cutting-edge technologies that drive our product forward.
About Parafin
Parafin is a leading technology firm dedicated to providing innovative infrastructure solutions. Based in San Francisco, we pride ourselves on our cutting-edge approach and commitment to excellence. Join us and be part of a team that values creativity and collaboration.
Similar jobs
1 - 20 of 7,725 Jobs
Search for Senior Software Engineer Data Infrastructure
At Plaid, we believe in the power of data-driven decision-making. Our data culture demands robust and scalable data systems that ensure accuracy and completeness. As a Senior Software Engineer focusing on Data Infrastructure, you will play a pivotal role in empowering teams across engineering, product, and business sectors to swiftly and securely extract valuable data insights. Your work will directly enhance our ability to serve customers effectively. You will be responsible for building and optimizing our data and machine learning infrastructure, allowing Plaid engineers to innovate and iterate on products built on consumer-permissioned financial data. Our Data Infrastructure engineers are experts in Data Warehousing, Data Lakehouse architecture, Spark, Workflow Orchestration, and Streaming technologies. You will enhance our existing data pipelines for performance and cost efficiency while creating intuitive abstractions that simplify the development process for other engineers at Plaid.
Founded in 2007, Airbnb has transformed the way people experience travel, connecting over 5 million hosts with more than 2 billion guests worldwide. Our platform enables unique stays and authentic experiences, fostering connections with local communities.The Team You Will Join:As a pivotal member of the Data Warehouse Infrastructure team, you will help shape the backbone of Airbnb's big data capabilities, enabling hundreds of engineers to efficiently collect, manage, and analyze vast amounts of data. We leverage cutting-edge open-source technologies such as Hadoop, Spark, Trino, Iceberg, and Airflow.Typical Responsibilities:Design and architect Airbnb's next-generation big data compute platform to enhance data ETL, analytics, and machine learning efforts.Oversee the platform's operations, focusing on improving reliability, performance, observability, and cost-effectiveness.Create high-quality, maintainable, and self-documenting code while engaging actively in code review processes.Contribute to open-source projects, making a significant impact on the industry.
Position OverviewJoin OpenEvidence as a Data Infrastructure Software Engineer, where you will engineer comprehensive systems that drive essential product and research operations. Your focus will be on optimizing performance, ensuring scalability, and enhancing accuracy, while enjoying the autonomy to manage the infrastructure that assists healthcare professionals in navigating complex clinical decisions in real-time.We value exceptional creators who thrive in versatile roles. Our engineers engage across various products and projects, taking ownership wherever they can make the most significant impact.About OpenEvidenceOpenEvidence is the leading medical AI platform globally, utilized by over 40% of clinicians in the U.S. in just over a year through organic product-led growth. As a $12 billion company, our engineering team comprises 30 talented individuals from MIT, Harvard, and Stanford. We believe that groundbreaking products are born from a small group of exceptional builders, driven by focused goals and empowered to take ownership and act swiftly. We are expanding our team to capitalize on an unparalleled opportunity to set the standard for medical AI platforms.If you are a top-tier engineer or scientist eager to push the boundaries and achieve tangible outcomes that affect millions of lives, we want to connect with you.Our CultureWe expect our work to be performed at an elite level. The journey from concept to execution and scaling is akin to a professional sport, where excellence is non-negotiable. We believe that the creation of innovative technologies is only achievable through complete ownership. Significant achievements happen when individuals take the initiative to see them through.Your ProfileThis role is not for those seeking a 9-to-5 job or merely looking to write papers. If you are ready to dive into the trenches, tackle challenges head-on, and create something from scratch that could impact millions and drive substantial revenue, you might be the perfect fit.We seek brilliant builders who are intelligent, ambitious, resourceful, self-reliant, detail-oriented, driven, hardworking, and humble. Does this sound rare? It is, as we have only found 30 of them so far, and we are eager to discover more.
Full-time|$200K/yr - $275K/yr|On-site|San Francisco
About Watney RoboticsAt Watney Robotics, we are pioneers in developing autonomous robotic solutions aimed at enhancing critical infrastructure. Recently securing $21 million in seed funding from leading investors such as Conviction, Abstract, and A*, we are collaborating with the world’s largest hyperscalers to propel the expansion of data centers and streamline maintenance processes.This is an extraordinary opportunity to join our team at a pivotal stage as we transition from prototype to large-scale production. Be part of a team that not only ships cutting-edge systems but also plays a crucial role in shaping the operational framework of an innovative robotics company.
Who We AreServal is an innovative AI-driven automation platform redefining operational efficiency for enterprises. Our intelligent agents seamlessly comprehend and execute real-world workflows, replacing outdated manual processes with adaptive, self-learning software. Since our inception in early 2024, we have garnered the trust of industry leaders such as General Motors, Notion, Perplexity, Vercel, Mercor, LangChain, and Verkada, streamlining high-volume operational tasks across their organizations.At the heart of Serval is a cutting-edge agentic AI platform that transforms natural language into actionable workflows. Our agents not only respond to queries but also reason, act across various systems, and continuously enhance their performance. What started as a solution for operational tasks has rapidly expanded into a versatile AI automation layer utilized across IT, HR, Finance, Security, Legal, and Engineering sectors.Our mission is to eradicate repetitive, manual tasks within enterprises, empowering teams through intelligent automation. In the long run, we aim to establish a universal AI operations layer—a system of agents that integrates across business functions, maintaining the momentum of modern companies.We are proud to be backed by renowned investors including Sequoia Capital, Redpoint Ventures, Meritech, First Round, General Catalyst, and Elad Gil, and founded by seasoned product and engineering leaders from Verkada.Role OverviewAs a Senior Software Engineer in Infrastructure at Serval, you will be pivotal in developing and scaling the core systems that empower our AI agents and workflow automation platform. A crucial aspect of this role involves enabling and supporting self-hosted deployments for enterprise clients needing on-premises or private cloud environments. We are looking for engineers with profound expertise in distributed systems, infrastructure-as-code, production operations, and customer-facing support, who aspire to influence the technical architecture of a rapidly evolving platform.What You'll DoDesign, implement, and operate large-scale distributed systems that power Serval's AI agents, workflow orchestration, and data pipelines.Create and maintain Terraform modules to provision and manage cloud infrastructure across AWS, GCP, or Azure environments.Develop and sustain deployment packages, installation scripts, and infrastructure templates, enabling customers to self-host Serval in their own environments.Provide technical support and guidance to enterprise customers during installation and deployment phases.
At Hover, we empower individuals to conceptualize, enhance, and safeguard the spaces they cherish. Utilizing proprietary AI and over a decade's worth of real property data, we provide answers to pivotal questions such as, 'What will it look like?' and 'What will it cost?' Our platform offers homeowners, contractors, and insurance professionals accurately measured, interactive 3D models of properties — all achievable from a smartphone scan in mere minutes.Driven by curiosity and purpose, we maintain a strong commitment to our customers, communities, and one another. We believe that diverse perspectives foster the best ideas, and we take pride in nurturing an inclusive, high-performance culture that encourages growth, accountability, and excellence. Supported by premier investors like Google Ventures and Menlo Ventures, and trusted by industry leaders such as Travelers, State Farm, and Nationwide, we are revolutionizing how individuals perceive and interact with their environments.About the RoleAs a Senior Software Engineer specializing in Infrastructure, you will delve into cloud infrastructure challenges unique to a company focused on 3D data, computer vision, and machine learning. Your enthusiasm for building internal tools and your talent for crafting elegant solutions to complex issues will be crucial in this role.Our Infrastructure team is responsible for everything beyond the application binary, serving as a critical partner to the rest of the engineering department. Through automation, we aim to streamline processes, ensuring that the simplest path is also the fastest and most secure. We manage and optimize all cloud infrastructure components including our Kubernetes environment, databases, networks, storage, and caching systems. Collaborating with engineering peers, we establish consistent solutions to common architectural challenges, particularly those involving rich geospatial and machine learning workloads. We are well-versed in best practices for cloud architecture and CI/CD, leveraging application development as a means to implement these practices.Your ContributionsYou will play a pivotal role in developing straightforward solutions to intriguing problems, thereby enhancing the foundation upon which our engineering teams build. Collaborating closely with engineers across the organization, you will help make their applications faster, easier to manage, and more reliable in production. Your work will span frontend, backend, computer vision, data, security, and machine learning teams to scale new ideas into production effectively. Given the small and highly collaborative nature of our team, you can expect a varied and impactful workload, which may include:Designing scalable cloud architectureEnhancing CI/CD pipelines and developer tooling
Full-time|$153K/yr - $376K/yr|Remote|San Francisco, CA • New York, NY • United States
At Figma, we are expanding our team of dedicated creatives and innovators committed to making design accessible for everyone. Our platform empowers teams to transform ideas into reality—whether you're brainstorming, prototyping, converting designs into code, or utilizing AI for enhancements. From concept to product, Figma enables teams to optimize workflows, accelerate processes, and collaborate in real-time from anywhere in the world. If you're passionate about shaping the future of design and teamwork, we invite you to join us!The Data Platform team at Figma is responsible for constructing and managing the essential systems that drive analytics, AI/ML initiatives, and data-informed decision-making across our organization. We cater to a wide array of stakeholders, including AI researchers, machine learning engineers, data scientists, product engineers, and business teams that depend on data for insights and strategic planning. Our team is tasked with owning and scaling critical platforms such as the Snowflake data warehouse, ML Datalake, orchestration and pipeline infrastructure, and extensive data ingestion and processing systems, overseeing all data transactions that occur within these platforms.Despite our small size, we tackle significant, high-impact challenges. In the upcoming years, we are focused on developing the data infrastructure layer for Figma's AI-driven products, enhancing cost and performance efficiencies across our data stack, scaling our ingestion and reverse ETL capabilities for new product applications, and reinforcing data quality, reliability, and compliance at every level. If you are enthusiastic about creating scalable, high-performance data platforms that empower teams across Figma, we would love to connect with you!This is a full-time role that can be performed from one of our US hubs or remotely within the United States.
Full-time|$350K/yr - $475K/yr|On-site|San Francisco
At Thinking Machines Lab, our vision is to enhance human potential by advancing collaborative general intelligence. We are dedicated to creating a future where individuals have the resources and knowledge to harness AI for their specific objectives and aspirations.Our team comprises scientists, engineers, and innovators who have developed some of the most popular AI products, including ChatGPT and Character.ai, as well as influential open-weight models like Mistral, along with highly regarded open-source projects such as PyTorch, OpenAI Gym, Fairseq, and Segment Anything.About the RoleWe are seeking a talented engineer to enhance our data infrastructure. You will become part of a dynamic, high-impact team tasked with designing and scaling the foundational infrastructure for distributed training pipelines, multimodal data catalogs, and sophisticated processing systems that manage petabytes of data.Our infrastructure is pivotal; it serves as the foundation for every groundbreaking achievement. You will collaborate directly with researchers to expedite experiments, develop novel datasets, optimize infrastructure efficiency, and derive essential insights from our data repositories.If you are passionate about distributed systems, large-scale data mining, and open-source tools such as Spark, Kafka, Beam, Ray, and Delta Lake, and enjoy building innovative solutions from scratch, we encourage you to apply.Note: This is an evergreen role that we keep open continuously for expressions of interest. We receive a high volume of applications, and while there may not always be an immediate position that aligns perfectly with your skills and experience, we encourage you to apply. We regularly review applications and reach out as new opportunities arise. You are welcome to reapply after gaining more experience, but please refrain from applying more than once every six months. We may also post for specific roles for particular projects or team needs, and in those cases, you are welcome to apply directly in addition to this evergreen role.
About Our TeamAt OpenAI, our Data Platform team is at the heart of our innovative approaches to data management, powering essential product, research, and analytics workflows. We manage some of the largest Spark compute fleets in production, architect data lakes and metadata systems on Iceberg and Delta, and envision exabyte-scale architectures. Our high-throughput streaming platforms utilize Kafka and Flink, while our orchestration is powered by Airflow. We also support machine learning feature engineering tools such as Chronon. Our mission is to provide secure, reliable, and efficient data access at scale, thereby enhancing intelligent, AI-assisted data workflows.Join us in building and maintaining these core platforms that are foundational to OpenAI's products, research, and analytics capabilities.We are not just scaling infrastructure; we are transforming the way people engage with data. Our vision includes intelligent interfaces and AI-powered workflows that make data interactions faster, more reliable, and intuitive.About the PositionIn this role, you will focus on constructing and managing data infrastructure that supports extensive compute fleets and storage systems optimized for high performance and scalability. You will be instrumental in designing, developing, and operating the next generation of data infrastructure at OpenAI. Your responsibilities will encompass scaling and securing big data compute and storage platforms, building and maintaining high-throughput streaming systems, ensuring low-latency data ingestion, and facilitating secure, governed data access for machine learning and analytics. You will also prioritize reliability and performance at extreme scales.You will have complete ownership of the full lifecycle: from architecture to implementation, production operations, and on-call responsibilities.You should be experienced with platforms such as Spark, Kafka, Flink, Airflow, Trino, or Iceberg. Familiarity with infrastructure tools like Terraform, along with expertise in debugging large-scale distributed systems, is essential. A passion for addressing data infrastructure challenges in the AI domain is a must.This role is based in San Francisco, CA. We offer a hybrid work model requiring 3 days in the office each week and provide relocation assistance for new hires.Responsibilities:Design, build, and maintain data infrastructure systems including distributed compute, data orchestration, distributed storage, streaming infrastructure, and machine learning infrastructure, ensuring they are scalable, reliable, and secure.Ensure our data platform can scale significantly while maintaining reliability and efficiency.Enhance company productivity by empowering your fellow engineers and teammates through innovative data solutions.
Full-time|$166K/yr - $225K/yr|On-site|San Francisco, California
P-78 While candidates in the listed locations are encouraged for this role, candidates in other locations will be considered. At Databricks, we are dedicated to empowering data teams to tackle the world's most challenging problems—from realizing the next mode of transportation to advancing medical breakthroughs. We accomplish this by creating and managing the premier data and AI infrastructure platform, enabling customers to leverage deep data insights for business enhancement. Founded by engineers and with a strong customer focus, we eagerly embrace every opportunity to address technical challenges, from crafting cutting-edge UI/UX for data interaction to scaling our services and infrastructure across millions of virtual machines. And this is just the beginning. As a Senior Software Engineer on the Infrastructure teams, you will develop scalable systems that underpin the Databricks platform, positioning it as the go-to solution for executing Big Data and AI workloads. Your role will involve enhancing the Databricks infrastructure platform, encompassing multi-cloud systems and services designed to manage thousands of Kubernetes clusters at scale, storing petabytes of data, providing highly scalable and distributed API gateways, implementing a rate limiting framework, ensuring network security and encryption, and creating developer tools and infrastructure (we utilize Bazel), testing frameworks, and scalable CI/CD systems, among many other responsibilities. The impact you will have: Expand and enhance key components of the core Databricks infrastructure. Design multi-cloud systems and abstractions to enable the Databricks product to operate across existing Cloud providers. Enhance software development workflows to improve engineering and operational efficiency. Utilize our own data and AI platform to analyze build and test logs and metrics, identifying areas for enhancement. Create automated build, test, and release infrastructures. Establish and maintain engineering process standards to support our growth and success.
Foxglove develops data infrastructure for robotics teams operating in real-world environments such as factories and warehouses. As robots leave the lab, engineers need reliable tools for analyzing data, diagnosing issues, and improving system performance. Foxglove delivers observability, visualization, and data management solutions designed to help teams manage large volumes of multimodal sensor data from deployed fleets. Role overview This Software Engineer - Robotics Data Infrastructure position centers on building and optimizing the systems behind Foxglove’s products. The scope covers desktop and web visualization tools, backend services for data ingestion and streaming, and client libraries running directly on robots. Work ranges from enhancing decoding performance in Rust, to extending MCAP tooling in C++, integrating new data sources with TypeScript, and occasionally working with customers to resolve performance issues. What you will do Design, build, and deploy product features from start to finish, incorporating feedback from users. Work across the stack: from Rust and C++ libraries on devices, to backend cloud services, to browser-based visualization tools. Identify and address performance bottlenecks in data pipelines, including ingestion, decoding, streaming, and rendering. Contribute to MCAP and other open-source libraries used by the robotics community. Collaborate with customers and robotics engineers to gather requirements and validate new solutions. Maintain high engineering standards and help foster a culture of ownership within the team. Design systems for efficient storage and querying of petabyte-scale robotics data. Requirements At least 5 years of experience developing production software. Strong proficiency in Rust, C++, and TypeScript, with a willingness to learn new languages or frameworks as needed. Location This position is based in San Francisco, CA.
Full-time|$160K/yr - $225K/yr|Hybrid|San Francisco, CA (Hybrid)
About Fable SecurityAt Fable Security, we recognize that AI-driven threats and human error pose significant risks to enterprise security. Cybercriminals exploit human behavior, which is responsible for over 70% of security breaches. Our mission is to empower individuals with the right tools, transforming them from targets into an active line of defense.We have developed a human risk platform that effectively shapes employee behavior. Our user-friendly and scalable platform integrates complex employee data, identifies risky behaviors, and automatically delivers timely, relevant interventions where employees are most engaged—in real time.Supported by renowned investors such as Redpoint Ventures and Greylock Partners, and founded by members of the Abnormal Security team, Fable is addressing one of cybersecurity’s most pressing challenges within a multi-billion-dollar market. Our diverse team includes alumni from Meta, Twitter, and prestigious universities like Columbia, Stanford, and UCLA. As we experience rapid growth, this is a prime opportunity to contribute to and influence the future of security.Why Join UsHelp us build and scale the core data infrastructure that drives a groundbreaking product.Collaborate with engineering, data science, and product teams to operationalize data effectively at scale.Be part of a small, elite team where your contributions will have a significant impact.As part of an early-stage company, every engineer plays a crucial role in shaping product functionality and evolution. You will define not only the technical architecture but also the company’s data philosophy.Your RoleIn the position of Data Infrastructure Engineer, you will be responsible for the architecture, scalability, and reliability of our data platform.You will design and construct systems that support everything from real-time product functionalities to internal analytics and machine learning processes, covering the spectrum from data ingestion to production-ready datasets. Additionally, you will establish best practices that underpin our data-driven products.This role is highly cross-functional, requiring close collaboration with engineering, data, and product teams to ensure our data foundation evolves in tandem with our growth.ResponsibilitiesDesign, develop, and sustain scalable data systems.Implement best practices for data architecture and management.Collaborate with cross-functional teams to facilitate data-driven decision-making.
Innovating the Future of SoftwareAs we approach 2026, the software industry is facing an unprecedented challenge: the 'infinite software crisis.' At Sazabi, we are dedicated to redefining how engineering teams support, maintain, and operate the rapid growth in application development.Introducing Sazabi: The AI-Native Observability Platform for Agile Engineering Teams.Our platform empowers teams by providing a centralized solution to inquire about their production systems in natural language, visualize system activities automatically, and diagnose issues ten times faster.Say goodbye to tedious instrumentation, dashboard setups, and alert tuning—just straightforward answers.We are proud to be backed by pioneers from leading AI organizations, including Vercel, Graphite, Daytona, Browserbase, LangChain, Mastra, Replit, and others.
Full-time|$162K/yr - $216K/yr|Hybrid|San Francisco, California, United States
Who We AreBaton is Ryder’s innovative product development division dedicated to leveraging cutting-edge technologies to transform the transportation and logistics landscape. Managing over $10 billion in freight, our technology has a significant impact across the U.S. economy.We are committed to creating and delivering software that not only meets but exceeds the needs of Ryder and its 50,000+ clients, which includes some of the most recognized brands globally. Our projects range from user-centric applications to the robust data platform that will drive the future of Ryder’s innovations.Baton’s mission: To enable a supply chain that operates on autopilot.Since Ryder’s acquisition of Baton in 2022, we have been operating with the agility of a startup while benefiting from the extensive reach of a Fortune 500 company. If you're passionate about tackling intricate challenges and making a real impact in the backbone of the American economy, you’ll thrive with us.Role: Software Engineer - InfrastructureDepartment: Data PlatformLocation: Hayes Valley, San Francisco, CA
Join our innovative team at Astranis as a Senior Software Engineer specializing in Infrastructure. In this role, you will be responsible for designing, implementing, and maintaining robust infrastructure solutions that support our cutting-edge satellite technology. Your expertise will play a crucial role in enhancing the reliability and scalability of our systems.
Full-time|$190K/yr - $280K/yr|Hybrid|San Francisco, California
About SentrySentry is dedicated to eliminating poor software experiences. Our mission is to empower developers to create high-quality software swiftly, allowing everyone to enjoy technology to its fullest.With over $217 million raised in funding and a community of over 100,000 organizations, including giants like Disney, Microsoft, and Atlassian, we are developing state-of-the-art performance and error monitoring tools. Our solutions help our partners minimize time spent on bug fixes and maximize product development.In our commitment to collaboration, Sentry follows a hybrid work model across our global offices. We have designated Mondays, Tuesdays, and Thursdays as in-office days to foster effective teamwork. If you are passionate about building tools that enhance the digital experience, join us in creating the next generation of software monitoring solutions.About the RoleAt Sentry.io, we offer vital services for diagnosing application health issues. Our tools are crucial for organizations aiming to respond adeptly in dynamic markets. We ensure a seamless and enjoyable experience in the development and deployment of these tools through a robust continuous integration environment and an insightful deployment pipeline.As part of the Infrastructure Engineering team, your contributions will be instrumental in supporting Sentry's growth and enabling engineering teams to operate with agility and confidence.Your responsibilities will include designing, developing, and maintaining internal software and platform capabilities that alleviate the cognitive load associated with infrastructure and developer tooling. You will create dependable, reusable abstractions that facilitate rapid shipping of features while incorporating durability, security, and operational excellence into service development and management.This role demands strong engineering judgment: selecting reliable technologies, planning for scalability from the outset, and crafting solutions that serve multiple teams. Your focus will be on practical systems that enhance reliability and ownership across the organization, driving adoption through comprehensive documentation, well-designed APIs, and seamless developer experiences that integrate into daily workflows.Ultimately, you will empower engineering teams to flourish within a culture of ownership—enabling them to deploy, manage, and evolve services confidently while minimizing operational burdens.Key ResponsibilitiesDesign systems that scale with company growth, ensuring a balance of reliability, performance, and cost-efficiency.Develop platform services that enhance internal operations and developer productivity.
Compensation: Competitive base salary + substantial equityBenefits: Health & dental insurance, gym reimbursement, daily team lunches, 401(K)About JuliusAt Julius, we're pioneering advancements in applied AI by developing cutting-edge coding agents. Our platform executes approximately 1 million lines of code every 36 hours, serving over 1 million users and generating 3 million+ visualizations. We manage all code in isolated remote containers. As a revenue-generating entity, we are backed by AI Grant and founders with remarkable backgrounds from companies like Vercel, Notion, Perplexity, Palantir, Replit, Zapier, Intercom, and Dropbox.The RoleJoin us in building and scaling the robust code-execution platform that powers Julius, across both cloud and on-prem environments. We orchestrate over 500,000 containers/month and the demand is growing rapidly. You will take ownership of reliability, performance, and security within our multi-tenant compute environment.Your ResponsibilitiesDesign and manage a secure, multi-tenant container infrastructure that ensures quick startup and intelligent autoscaling.Implement on-prem/private cloud deployments using Helm and Terraform, integrating SSO, network controls, and audit logging.Enhance observability (metrics, traces, logs) with well-defined SLOs and lead incident response initiatives.Optimize images, scheduling, networking, and costs, while developing fair-use and rate-limiting controls.Your QualificationsStrong experience with production Kubernetes and container internals (Docker/containerd); solid understanding of networking principles.Familiarity with cloud environments (AWS/GCP/Azure) and Infrastructure as Code (Terraform/Helm).Proficiency in monitoring and logging tools (Prometheus, Grafana, OpenTelemetry, ELK/Vector).Understanding of security best practices for containerized, multi-tenant systems.Preferred QualificationsExperience with gVisor, Kata, Firecracker; Cilium/eBPF; GPU scheduling; serverless autoscaling (KEDA/Knative/Karpenter).Proven experience delivering on-prem or air-gapped enterprise software solutions.A passion for AI, with experience building side projects involving LLMs.Why Join Julius?Be part of a small, senior team where your contributions will have a massive impact. Tackle challenging infrastructure problems at a meaningful scale.
At Judgment Labs, we specialize in developing cutting-edge infrastructure for Agent Behavior Monitoring (ABM). Unlike conventional observability tools that merely track exceptions and latency, our ABM technology identifies behavioral anomalies, such as instruction drifts and context retrieval losses, in large-scale production settings.Our solutions are trusted by numerous teams working on autonomous agents to gain insights into system behavior post-deployment. Rather than simply reacting to incidents, our clients analyze patterns across conversations and workflows, correlate regressions with specific interaction types, and identify critical points of reliability failure. Recently, we secured over $30 million across two funding rounds from notable investors like Lightspeed, SV Angel, and Valor Equity Partners.The Role:We are seeking a Senior Data Infrastructure Engineer to architect and enhance the real-time data pipelines essential for robust agent behavior analysis at scale. This position plays a vital role in processing hundreds of thousands of traces per second, executing LLM-based scoring and clustering in near-real-time, and ensuring low-latency query performance, which allows teams to monitor agent behavior as it unfolds. Ideal candidates will have experience designing petabyte-scale data systems, optimizing OLAP database performance, and managing the full data lifecycle from ingestion to analytics.What You'll Do:Design and automate large-scale, high-performance streaming and batch data processing systems to support Judgment's behavioral analysis products.Collaborate closely with infrastructure and backend teams to enhance scalability, data governance, and operational efficiency.Promote best practices in software engineering for data infrastructure at scale.Uphold high standards for data quality and engineering: ensuring reliability, efficiency, documentation, testability, and maintainability.Craft data models for optimal storage and access, ensuring efficient data flows to meet critical product requirements.Enhance OLAP database performance through careful schema design, partitioning strategies, storage optimization, and access pattern analysis.
Senior Software Engineer, Infrastructure & PlatformRole OverviewIn the role of Senior Software Engineer, Infrastructure & Platform at AfterQuery, you will take on the exciting challenge of designing and constructing the essential infrastructure that drives our innovative data generation, evaluation, and agentic systems.Your responsibilities will include developing shared platforms that empower our engineering and research teams to execute large-scale human-in-the-loop workflows, evaluation harnesses, and automated data pipelines essential for training cutting-edge AI models.This position demands a high level of technical expertise and offers extensive ownership. You will be responsible for architecting and building the foundational infrastructure relied upon by numerous engineers, ensuring that systems are scalable, reliable, and capable of handling high-throughput workloads.Collaboration with the founding team will be key as you define system architecture, establish best engineering practices, and create the infrastructure that supports the evolution of AI development.
Join our dynamic team at Parafin as a Senior Software Engineer specializing in Infrastructure. In this pivotal role, you will design, develop, and maintain robust infrastructure solutions that support our scalable applications. Your expertise will help us enhance system performance, reliability, and security.We are looking for innovative thinkers who thrive in a collaborative environment. You will work closely with cross-functional teams to implement cutting-edge technologies that drive our product forward.
Apr 3, 2026
Sign in to browse more jobs
Create account — see all 7,725 results
Tailoring 0 resumes…
Tailoring 0 resumes…
We'll move completed jobs to Ready to Apply automatically.