Senior Systems Software Engineer
Clicking Apply Now takes you to AutoApply where you can tailor your resume and apply.
Unlock Your Potential
Generate Job-Optimized Resume
One Click And Our AI Optimizes Your Resume to Match The Job Description.
Is Your Resume Optimized For This Role?
Find Out If You're Highlighting The Right Skills And Fix What's Missing
Experience Level
Senior
Qualifications
About Lumafield
Lumafield is an innovative company at the forefront of engineering technology, committed to revolutionizing how engineers work with cutting-edge X-ray CT scanning solutions. Our mission is to provide accessible tools that enhance product visibility and drive efficiency in engineering processes.
Similar jobs
Search for Software Engineer Ray Data
6,189 results
Anyscale
About the RoleAnyscale is actively seeking a Senior Product Manager to spearhead the development of Ray Data, our innovative and scalable data processing library tailored for Machine Learning (ML) and Artificial Intelligence (AI) workloads.This position presents a unique opportunity to navigate the challenges of merging open-source growth with commercial innovation. You will drive rapid adoption within the open-source Ray Data ecosystem while simultaneously crafting powerful proprietary features for RayTurbo Data, our advanced commercial engine.As the steward of the Ray Data product roadmap, you will collaborate closely with engineering teams, sales, enterprise clients, and the open-source community. Success in this role hinges on your ability to:Immerse yourself in the end-user experience to identify product gaps and trade-offsEngage with customers and open-source users to expertly balance growth and commercializationStrategically assess the ML/Data lifecycle, pinpointing where our architectural strengths can deliver maximum valueArticulate a clear and compelling product strategy to stakeholdersKey ResponsibilitiesLead the Ray Data product strategy - Harmonize development of open-source features with the distinctive capabilities of RayTurbo Data to solidify Ray Data's position as the open-source standard for AI data processing.Enhance open-source Ray Data adoption - Prioritize community engagement, developer experience, and ecosystem partnershipsMarket Positioning & Enablement - Collaborate with Product Marketing on strategic positioning, enablement for sales teams, and conducting competitive analysis to ensure ongoing differentiation.Customer Engagement - Foster critical customer relationships to support sales and field engineering efforts.Required Qualifications3+ years of experience in product management, specifically with technical productsRobust technical expertise in distributed systems, ML infrastructure, or data processingProven experience engaging with both developer and enterprise audiencesStrong strategic thinking skills with the ability to manage competing priorities and stakeholder expectationsExcellent communication skills for articulating vision and strategy
Lumafield
About Lumafield: Established in 2019, Lumafield has pioneered the development of the world's first accessible X-Ray CT scanner specifically designed for engineers. Our intuitive scanner, combined with cloud-based software, empowers engineers to gain unparalleled insights into their projects at a remarkably affordable cost. Engineers face high-stakes decisions daily, necessitating tools that provide maximum visibility into their designs. By delivering exceptional product clarity and AI-enhanced tools that identify issues and produce quantitative insights, Lumafield is set to transform the creation, manufacturing, and application of complex products across various sectors. Our company thrives on impact and is dedicated to delivering the utmost value to our customers, ensuring their needs drive our development. Our talented team consists of leading researchers, industrial designers, PhD holders, innovators, and startup founders, all working collaboratively without egos. We proudly receive backing from prestigious venture capital firms, including Kleiner Perkins, Lux Capital, DCVC, and Spark Capital.Headquartered in Cambridge, MA, with an additional office in San Francisco, CA, we are excited to grow our team.About the Role: As a Senior Systems Software Engineer at Lumafield, you will be instrumental in developing the software that drives our cutting-edge, in-line manufacturing CT scanning products. You will engage with state-of-the-art X-ray physics, high-speed detectors, image processing, and embedded systems. Collaborating within a small team focused on our latest hardware, you will harness your expertise to maximize system performance and achieve outstanding results for our clients. This position is perfect for those eager to take ownership of embedded systems, firmware, and software design in an early-stage product environment. This role is based in our San Francisco, CA office, with occasional travel required to our Cambridge, MA office.
About Our TeamAt DoorDash, data drives our success. Our Data Engineering team is pivotal in building robust database solutions tailored for diverse applications, including reporting, product analytics, marketing optimization, and financial reporting. By architecting pipelines, data structures, and data warehouse environments, we enable data-driven decision-making across the organization.About the RoleWe are seeking a talented Software Engineer II to join our team as a technical leader, responsible for scaling our data infrastructure, enhancing automation, and developing tools to support our expanding business needs.What You Will DoCollaborate with business partners and stakeholders to gather and understand data requirements.Work alongside engineering, product teams, and external partners to ensure seamless data collection.Design, develop, and implement high-performance data models and pipelines for our Data Lake and Data Warehouse.Establish and execute data quality checks, conduct thorough QA, and implement monitoring routines.Enhance the reliability and scalability of our ETL processes.Manage a suite of data products that deliver accurate and trustworthy data.Support and onboard new engineers as they join our team.What We Are Looking For3+ years of professional experience in data engineering, business intelligence, or a related field.Proficiency in programming languages such as Python and Java.3+ years of experience with ETL orchestration and workflow management tools, including Airflow, Flink, Oozie, and Azkaban, using AWS/GCP platforms.Strong understanding of database fundamentals, SQL, and distributed computing.3+ years of experience with distributed data ecosystems (e.g., Spark, Hive, Druid, Presto) and streaming technologies like Kafka and Flink.Experience with Snowflake, Redshift, PostgreSQL, and/or other database management systems.Excellent communication skills with a proven ability to liaise with both technical and non-technical teams.Familiarity with reporting tools such as Tableau, Superset, and Looker.Able to thrive in a fast-paced and dynamic environment.
Condor Software
About CondorAt Condor, we are transforming the financial infrastructure of clinical development. While substantial investments are made annually to discover and develop new therapies, the processes behind these advancements often remain outdated and disconnected. Our mission is to bridge this gap, creating a cohesive system that integrates clinical operations, vendor activities, and financial data into a real-time intelligence layer. This empowers R&D and finance teams with the insights they need to make informed decisions.Our AI-driven, pharma-native infrastructure is designed to scale industry standards that we have helped shape alongside major partners. We facilitate prediction, control, and execution in some of the most complex R&D environments globally.As we continue to gain the trust of enterprise teams, we are now focused on the critical task of scaling our operations in a high-stakes environment.Condor is a rapidly growing company, backed by leading institutional investors such as Felicis and 645 Ventures, collaborating with top 200 biopharma companies. This is a unique opportunity to contribute to the infrastructure that influences how new therapies reach patients.The RoleWe are seeking a Senior Backend and Data Platform Engineer to play a key role in developing the foundational data infrastructure for Condor’s financial intelligence platform. This position is pivotal in turning complex clinical and financial data into actionable intelligence that enterprise biopharma teams can rely on.In this role, you will be responsible for designing and managing the core data foundations that underpin Condor’s financial engine and AI capabilities. Your work will involve modeling intricate, high-stakes data, constructing reliable data pipelines and services, and ensuring that product features and intelligence workflows function with precision, consistency, and scalability. The systems you develop will directly support critical finance and operational applications.This hands-on, senior engineering position provides you with significant ownership. You will engage with backend services, data pipelines, and APIs, bringing features from concept to production. You will define necessary data schemas, transformations, and architectural patterns that become essential as our platform evolves. Although your primary focus will be on backend and data engineering, you will also be encouraged to work across the stack to ensure seamless integration of data and intelligence.
sfcompute
Role Overview sfcompute is hiring a Software Engineer focused on ETL and Data in San Francisco, CA. This position centers on building and maintaining data pipelines that turn raw data into actionable insights. What You Will Do Design and implement ETL processes to move and transform data efficiently Work with teams across the company to improve data quality and accessibility Support data-driven decision-making by ensuring reliable and accurate information is available
Airbnb, Inc.
Founded in 2007, Airbnb began its journey when two hosts welcomed three guests into their San Francisco home. Today, we boast a thriving community of over 5 million hosts who have welcomed more than 2 billion guest arrivals across nearly every country worldwide. Our hosts provide exceptional stays and unique experiences, enabling guests to connect with local communities in a genuine and meaningful way.Join Our Community:At Airbnb, we prioritize the importance of reliable data across all business sectors to drive insight and innovation. To achieve this, we focus on understanding business needs, securing appropriate data sources, designing effective data models, and establishing robust and dependable data pipelines.We are currently recruiting for the following teams:The Data Stewardship Team: A dedicated group of data enthusiasts with diverse expertise in analytics, data modeling, governance, compliance, and scalable data quality. Our mission is to ensure that Airbnb meets its compliance obligations within our data ecosystem while enabling data consumers to easily find the best data suited for their needs. As part of the overall Data Infrastructure organization, we manage the online and offline data infrastructure and oversee the processes that facilitate data transitions between these environments.The Users and Contextualization Data & AI Team: A crucial component of the Marketplace Data & AI, this team focuses on developing foundational data systems that provide deeper insights into essential domains. Specifically, we concentrate on user data (Guests & Hosts) to create high-quality, well-governed user data and insights. These insights are vital for crafting personalized and context-aware experiences that enhance trip quality both on and off the Airbnb platform, ultimately enabling Airbnb to better understand and serve its users throughout their journey.Your Impact:Data Stewardship: This is integral to Airbnb's operations. High-quality data is imperative for our business decisions and the future of our AI initiatives. We are responsible for the overall strategy regarding data quality, identifying critical data and its provenance, measuring the effectiveness of our internal data products, and collaborating with our core catalog team to provide optimal data solutions.
Join SoFi as a Senior Software Engineer in our Data Foundations team, where you will play a pivotal role in shaping our data architecture and enhancing our data-driven capabilities. You will work closely with cross-functional teams to develop robust data solutions that empower our business decisions and improve customer experiences.As a Senior Software Engineer, you will leverage your expertise in data engineering, software development, and cloud technologies to build scalable data pipelines and maintain high-quality data infrastructure. Your contributions will directly impact our ability to deliver innovative financial solutions.
Discord Inc.
Join Discord, a platform that connects over 200 million users monthly, primarily through gaming. With over 90% of our users engaging in gaming activities, they spend a staggering 1.5 billion hours playing unique titles every month. Discord is pivotal in shaping the future of gaming, enabling seamless interactions before, during, and after gameplay.Every day, countless gamers gather on Discord to strategize, celebrate victories, and foster communities around their favorite games. Behind our voice channels and streaming sessions lies vast amounts of data, capturing the essence of gaming connections.We are in search of an enthusiastic Software Engineer eager to construct data infrastructure at an enormous scale, with a genuine passion for supporting gaming communities. You will play a crucial role in developing the next-generation Data Platform that drives insightful decisions on one of the most vibrant platforms globally.If you are a Software Engineer who thrives on discussing distributed systems, prioritizes user privacy, and seeks to make data-driven decisions that impact millions of gamers worldwide, we would love to hear from you.For more insights into Discord's Data Platform, check out our engineering blog, including how we built our modern data stack leveraging open-source tools!
Probably Genetic
About Probably GeneticProbably Genetic is revolutionizing the lives of patients with severe and complex diseases. Our advanced data platform empowers drug developers and patient advocacy organizations to create and launch innovative treatments. By leveraging cutting-edge technology, we identify undiagnosed patients online, analyze their conditions using machine learning and home testing, and facilitate compliant communication with them. Our mission is to ensure that patients gain access to diagnoses, clinical trials, and treatments at the earliest opportunity.We are a dedicated team of passionate problem solvers, driven by a purpose that transcends individual interests. By prioritizing patient welfare, we are developing groundbreaking solutions in healthcare, with a roadmap full of innovations in bioinformatics, AI, and drug development. We invite you to join our lean, talented team and contribute to our vision.Probably Genetic has secured multiple funding rounds from top-tier Silicon Valley investors, including Threshold, Khosla, and Y Combinator. We offer competitive salaries, comprehensive benefits, and meaningful equity opportunities for early-stage team members.About the RoleWe are seeking a founding Data Engineer who is enthusiastic about shaping the future of data utilization to enhance patient outcomes. In this pivotal role, you will establish our data engineering architecture and construct the pipelines that drive internal insights and commercial data products. Your contributions will be instrumental in fostering clarity, impact, and growth throughout our organization.What You Will DoCollaborate closely with the Head of Engineering and Head of Product to transform complex data challenges into elegant, scalable solutions.Build reliable, maintainable infrastructure on AWS using Terraform to accommodate our expanding data requirements.Design data tables and pipelines tailored to the specific needs of our customers and internal teams.Implement state-of-the-art data pipelines with built-in observability from day one.Analyze and visualize data using BI tools to facilitate informed business decisions and provide customized insights to clients.Communicate your work and its impact across teams — presenting findings, receiving feedback, and continuously enhancing processes.Who You AreWe are eager to connect with candidates from diverse backgrounds who are committed to learning, growth, and making a meaningful impact. Here are a few attributes that will enable you to thrive in this role:Proficiency in data engineering principles and practices.Experience with cloud platforms, particularly AWS.Strong analytical skills and familiarity with data visualization tools.Ability to collaborate effectively in a team environment.
At Superhuman, we embrace a dynamic hybrid working model, allowing our team members to enjoy focused work time alongside in-person collaboration. This flexible approach fosters trust, innovation, and a vibrant team culture.We are seeking talented individuals based in San Francisco or Seattle for this role.About UsSuperhuman, an innovative AI productivity platform, aims to unlock the superhuman potential within everyone. We integrate AI seamlessly into the workflows of over 40 million people and 50,000 organizations worldwide, offering tools like Grammarly’s writing assistant, Coda’s collaborative workspaces, and our proactive AI assistant Go. Founded in 2009, we are committed to eliminating busywork, enabling our users to focus on what truly matters. Discover more at superhuman.com and learn about our core values here.Join Our TeamWe are on the lookout for a Senior Software Engineer to enhance our Data Enablement team and develop a world-class data platform. Your expertise will be crucial as we manage over 70 billion daily events, driving product enhancements through our systems. This role offers a unique opportunity to engage in every facet of complex software system development, from strategy formulation to architecture design and production deployment.Our engineers and researchers are encouraged to innovate and contribute to significant breakthroughs, shaping the trajectory of our product suite. As we scale our operations, we face increasingly complex technical challenges, providing fertile ground for growth and development. To gain further insights into our work, check out our technical blog.
Superhuman embraces a vibrant hybrid work model, offering team members a blend of focused work and collaborative in-person interactions that cultivate trust, innovation, and a robust company culture. Team members in this role should be located in San Francisco or New York City. About SuperhumanSuperhuman, home to Grammarly, is an innovative AI productivity platform dedicated to unlocking the extraordinary potential in everyone. Our suite of applications seamlessly integrates with over one million tools and websites, featuring Grammarly’s writing assistance, Coda’s collaborative workspaces, Mail’s inbox management, and Go, the proactive AI assistant that delivers contextual help automatically. Founded in 2009, Superhuman empowers over 40 million individuals, 50,000 organizations, and 3,000 educational institutions globally, helping them eliminate busywork and focus on what truly matters. Discover more at superhuman.com and learn about our values here.The OpportunityWe are on the lookout for a Senior Engineer to join our Data Platform team, tasked with constructing an exceptional data platform. Superhuman’s success hinges on our capacity to efficiently process 60 to 70 billion daily events, utilizing our systems to enhance our product. This role offers a unique chance to engage in all facets of developing complex software systems, including shaping strategy, defining architecture, and executing deployments to production.Our engineers and researchers enjoy the freedom to innovate and make impactful contributions to our product roadmap. As we scale our interfaces, algorithms, and infrastructure, the complexity of our technical challenges is rapidly increasing. Dive deeper into our team's insights on our technical blog.OverviewAs a Software Engineer on the Data Platform team, you will be instrumental in defining the architecture and technical strategy for our data platform, ensuring scalability, security, and efficiency across our data engineering systems. You will design and spearhead the implementation of robust, scalable, and dependable systems that manage high volumes of data, facilitating both product features and data-driven decision making across the organization. Your responsibilities will encompass areas such as real-time data processing and ETL (Extract, Transform, Load).
OpenEvidence
Position OverviewJoin OpenEvidence as a Data Infrastructure Software Engineer, where you will engineer comprehensive systems that drive essential product and research operations. Your focus will be on optimizing performance, ensuring scalability, and enhancing accuracy, while enjoying the autonomy to manage the infrastructure that assists healthcare professionals in navigating complex clinical decisions in real-time.We value exceptional creators who thrive in versatile roles. Our engineers engage across various products and projects, taking ownership wherever they can make the most significant impact.About OpenEvidenceOpenEvidence is the leading medical AI platform globally, utilized by over 40% of clinicians in the U.S. in just over a year through organic product-led growth. As a $12 billion company, our engineering team comprises 30 talented individuals from MIT, Harvard, and Stanford. We believe that groundbreaking products are born from a small group of exceptional builders, driven by focused goals and empowered to take ownership and act swiftly. We are expanding our team to capitalize on an unparalleled opportunity to set the standard for medical AI platforms.If you are a top-tier engineer or scientist eager to push the boundaries and achieve tangible outcomes that affect millions of lives, we want to connect with you.Our CultureWe expect our work to be performed at an elite level. The journey from concept to execution and scaling is akin to a professional sport, where excellence is non-negotiable. We believe that the creation of innovative technologies is only achievable through complete ownership. Significant achievements happen when individuals take the initiative to see them through.Your ProfileThis role is not for those seeking a 9-to-5 job or merely looking to write papers. If you are ready to dive into the trenches, tackle challenges head-on, and create something from scratch that could impact millions and drive substantial revenue, you might be the perfect fit.We seek brilliant builders who are intelligent, ambitious, resourceful, self-reliant, detail-oriented, driven, hardworking, and humble. Does this sound rare? It is, as we have only found 30 of them so far, and we are eager to discover more.
Airbnb, Inc.
Founded in 2007, Airbnb has transformed the way people experience travel, connecting over 5 million hosts with more than 2 billion guests worldwide. Our platform enables unique stays and authentic experiences, fostering connections with local communities.The Team You Will Join:As a pivotal member of the Data Warehouse Infrastructure team, you will help shape the backbone of Airbnb's big data capabilities, enabling hundreds of engineers to efficiently collect, manage, and analyze vast amounts of data. We leverage cutting-edge open-source technologies such as Hadoop, Spark, Trino, Iceberg, and Airflow.Typical Responsibilities:Design and architect Airbnb's next-generation big data compute platform to enhance data ETL, analytics, and machine learning efforts.Oversee the platform's operations, focusing on improving reliability, performance, observability, and cost-effectiveness.Create high-quality, maintainable, and self-documenting code while engaging actively in code review processes.Contribute to open-source projects, making a significant impact on the industry.
At Plaid, we believe in the power of data-driven decision-making. Our data culture demands robust and scalable data systems that ensure accuracy and completeness. As a Senior Software Engineer focusing on Data Infrastructure, you will play a pivotal role in empowering teams across engineering, product, and business sectors to swiftly and securely extract valuable data insights. Your work will directly enhance our ability to serve customers effectively. You will be responsible for building and optimizing our data and machine learning infrastructure, allowing Plaid engineers to innovate and iterate on products built on consumer-permissioned financial data. Our Data Infrastructure engineers are experts in Data Warehousing, Data Lakehouse architecture, Spark, Workflow Orchestration, and Streaming technologies. You will enhance our existing data pipelines for performance and cost efficiency while creating intuitive abstractions that simplify the development process for other engineers at Plaid.
Jordan Park Group
Jordan Park Group is a premier investment management firm dedicated to providing tailored financial advice to a unique clientele of individuals, families, and institutions. Our clients are distinguished leaders in business, philanthropy, government, and society. Our diverse team embodies empathy and expertise, tackling complexity and challenges to empower clients in achieving their financial aspirations.Our mission is to enhance lives and legacies through strategic financial management.Associate Software Engineer - Data PlatformSan Francisco, CaliforniaYour OpportunityWe are looking for a motivated Technology & Data Engineering Associate to elevate our data platform and enhance data quality initiatives. This role sits at the dynamic intersection of data engineering, analytics, data quality management, and AI integration.The Associate will be responsible for designing, building, and maintaining production-grade data systems, collaborating closely with executives, investment teams, operations, and client services to achieve significant business outcomes. This is a hands-on technical position that offers substantial ownership and visibility within the organization.Your RoleData PlatformDesign, build, and maintain robust SQL data pipelines (e.g., dbt + Snowflake + Python)Develop and manage orchestration workflows (Airflow or equivalent)Create data ingestion processes and API integrationsMonitor pipeline reliability and performance; proactively resolve issuesData Quality & ControlsDevelop deep expertise in the firm’s data, systems, and processesDesign reconciliation frameworks across systems and automate data validation and exception reportingPerform detailed variance analysis and root cause investigationsCollaborate with operations and finance teams to ensure reporting accuracyAnalytics & Business IntelligenceArchitect and maintain Tableau dashboards utilized across the firmTrain and support internal stakeholders on self-service analyticsPython Development & AutomationDevelop Python-based tools for API integrations, data transformations, and ad-hoc analysis
Pinterest, Inc.
Join Pinterest as a Staff Software Engineer specializing in Conversion Data Privacy. In this pivotal role, you will lead the development of innovative solutions that enhance user privacy while supporting our conversion processes. You will collaborate with cross-functional teams to design, implement, and maintain systems that ensure compliance with data privacy regulations and foster user trust.As a key member of our engineering team, you will leverage your expertise in software development and data management to drive projects that have a significant impact on Pinterest's growth and user engagement.
Ripple Labs Inc.
At Ripple, we are on a mission to transform the way value is exchanged globally, making it as seamless as information transfer. Our innovative crypto solutions empower financial institutions, businesses, governments, and developers, promoting a more equitable financial system while creating opportunities for individuals across the globe. Joining us means being part of an impactful journey where you can hone your skills and collaborate with a supportive team.If you are eager to make a significant impact and explore exciting career advancement opportunities, we invite you to join us in building real-world value.THE WORK: WHAT YOU’LL DO: WHAT YOU'LL BRING: Other common names for this role:
P-186 At Databricks, we are passionate about empowering data teams to tackle some of the world’s most challenging problems, from security threat detection to cancer drug development. Our mission is to build and operate the leading data and AI infrastructure platform, enabling our customers to concentrate on the high-value challenges that are integral to their own objectives. Founded in 2013 by the original creators of Apache Spark™, Databricks has rapidly evolved from a small office in Berkeley, California, to a global powerhouse with over 1000 employees. Trusted by thousands of organizations, from startups to Fortune 100 companies, we are recognized as one of the fastest-growing SaaS companies worldwide. Our engineering teams create highly sophisticated products that address significant needs in the industry. We continuously push the limits of data and AI technology while maintaining the resilience, security, and scalability essential for our customers' success on our platform. We manage one of the largest-scale software platforms, consisting of millions of virtual machines that generate terabytes of logs and process exabytes of data daily. At this scale, we frequently encounter cloud hardware, network, and operating system faults, and our software must effectively shield our customers from these challenges. Modern data analysis leverages advanced techniques, such as machine learning, that far exceed the capabilities of traditional SQL query engines. As a Software Engineer on the Runtime team at Databricks, you will be instrumental in developing the next generation of distributed data storage and processing systems that outshine specialized SQL query engines in relational query performance, while providing the flexibility and programming abstractions to support a variety of workloads, from ETL to data science. Examples of projects you may work on include: Apache Spark™: Contributing to the de facto open-source framework for big data. Data Plane Storage: Developing reliable, high-performance services and client libraries for storing and accessing vast amounts of data on cloud storage backends like AWS S3 and Azure Blob Store. Delta Lake: A storage management system that merges the scalability and cost-effectiveness of data lakes with the performance and reliability of data warehouses, featuring low latency streaming. Its higher-level abstractions and guarantees, including ACID transactions and time travel, significantly reduce the complexity of real-world data engineering architectures. Delta Pipelines: Aiming to simplify the management of data engineering pipelines.
At Figma, we are expanding our team of dedicated creatives and innovators committed to making design accessible for everyone. Our platform empowers teams to transform ideas into reality—whether you're brainstorming, prototyping, converting designs into code, or utilizing AI for enhancements. From concept to product, Figma enables teams to optimize workflows, accelerate processes, and collaborate in real-time from anywhere in the world. If you're passionate about shaping the future of design and teamwork, we invite you to join us!The Data Platform team at Figma is responsible for constructing and managing the essential systems that drive analytics, AI/ML initiatives, and data-informed decision-making across our organization. We cater to a wide array of stakeholders, including AI researchers, machine learning engineers, data scientists, product engineers, and business teams that depend on data for insights and strategic planning. Our team is tasked with owning and scaling critical platforms such as the Snowflake data warehouse, ML Datalake, orchestration and pipeline infrastructure, and extensive data ingestion and processing systems, overseeing all data transactions that occur within these platforms.Despite our small size, we tackle significant, high-impact challenges. In the upcoming years, we are focused on developing the data infrastructure layer for Figma's AI-driven products, enhancing cost and performance efficiencies across our data stack, scaling our ingestion and reverse ETL capabilities for new product applications, and reinforcing data quality, reliability, and compliance at every level. If you are enthusiastic about creating scalable, high-performance data platforms that empower teams across Figma, we would love to connect with you!This is a full-time role that can be performed from one of our US hubs or remotely within the United States.
Senior Software Engineer - Data AcquisitionOverview:Join the dynamic Data Acquisition team at OpenAI, where we spearhead the data collection efforts essential for powering our advanced model training operations. Our team plays a pivotal role in managing web crawling and GPTBot services, collaborating closely with Data Processing, Architecture, and Scaling teams. We are seeking a talented Senior Software Engineer to enhance our Data Acquisition initiatives.Key Responsibilities:Lead engineering projects focused on data acquisition, including web crawling, data ingestion, and search optimization.Collaborate effectively with cross-functional teams to maintain seamless data flow and system performance.Engage with the legal team to navigate compliance and data privacy regulations.Design and implement robust distributed systems capable of processing petabytes of data.Develop algorithms for efficient data indexing and search functionalities.Build and sustain backend services for data storage, including experience with key-value databases and data synchronization.Deploy solutions in a Kubernetes Infrastructure-as-Code environment while conducting regular system audits.Conduct experiments on data to derive insights that drive system enhancements.Qualifications:Bachelor's, Master's, or PhD in Computer Science or a related discipline.A minimum of 6 years of professional experience in software development.Prior experience with large-scale web crawlers is a significant advantage.In-depth knowledge of large stateful distributed systems and data processing techniques.Expertise in Kubernetes and familiarity with Infrastructure-as-Code practices.A proactive approach to exploring new technologies and methodologies.Strong ability to juggle multiple tasks and adapt to changing priorities.Excellent communication skills, both written and verbal.About OpenAI:OpenAI is at the forefront of artificial intelligence research and deployment, dedicated to ensuring that the benefits of general-purpose AI are shared by all of humanity. We strive to push the boundaries of innovation while adhering to ethical standards.
Sign in to browse more jobs
Create account — see all 6,189 results

