DatabricksMountain View, California; San Francisco, California
On-site Full-time
Clicking Apply Now takes you to AutoApply where you can tailor your resume and apply.
Unlock Your Potential
Generate Job-Optimized Resume
One Click And Our AI Optimizes Your Resume to Match The Job Description.
Is Your Resume Optimized For This Role?
Find Out If You're Highlighting The Right Skills And Fix What's Missing
Experience Level
Manager
Qualifications
Proven experience in engineering management, preferably in data engineering or related fields. Strong understanding of data pipeline architectures and technologies. Excellent leadership and communication skills, with the ability to inspire and motivate teams. Experience with cloud platforms and big data technologies. Ability to drive projects from conception to completion with a focus on quality and performance.
About the job
Databricks is seeking an experienced Engineering Manager to lead our Pipelines Engine team. In this role, you will oversee the development and optimization of our data pipeline infrastructure, ensuring that we deliver high-performance solutions that meet the needs of our clients. You will collaborate with cross-functional teams to drive innovation and maintain our leadership in the data analytics space.
About Databricks
At Databricks, we are at the forefront of data analytics and machine learning innovation. Our collaborative culture fosters creativity and empowers our employees to make a difference. Join us in our mission to simplify the complexities of big data and help organizations succeed through data-driven decision-making.
Full-time|On-site|Mountain View, California; San Francisco, California
Databricks is seeking an experienced Engineering Manager to lead our Pipelines Engine team. In this role, you will oversee the development and optimization of our data pipeline infrastructure, ensuring that we deliver high-performance solutions that meet the needs of our clients. You will collaborate with cross-functional teams to drive innovation and maintain our leadership in the data analytics space.
Full-time|$139K/yr - $223K/yr|On-site|San Francisco, California
About UsAt Aurora, we are on a mission to make self-driving technology safe, swift, and accessible to everyone.The Aurora Driver is set to usher in a new era of mobility and logistics, fostering a future that is not only safer but also more efficient and accessible. Joining Aurora means tackling complex challenges alongside a team of dedicated and talented individuals, enhancing your expertise while broadening your knowledge base. For the latest updates from Aurora, visit aurora.tech or connect with us on LinkedIn. At Aurora, we seek out talented individuals from diverse backgrounds eager to contribute to a transportation ecosystem that enhances road safety, ensures timely delivery of essential goods, and promotes efficient and accessible mobility for all. We are currently on the lookout for a Graphics Pipeline Engineer.Key Responsibilities:Lead the technical execution of cross-functional projects, translating stakeholder needs into robust code while exemplifying engineering best practices.Design and implement foundational Python frameworks, services, and APIs that underpin our synthetic data ecosystem. This is a hands-on role requiring frequent coding and deployment.Champion the adoption and standardization of USD as the foundational data backbone for our pipeline.Serve as the lead developer and subject matter expert for our most intricate pipeline challenges, troubleshooting complex technical issues and engineering scalable solutions.
Join Mithrl as a Lead Bioinformatics EngineerAt Mithrl, we envision a future where groundbreaking medicines are swiftly delivered to patients, transforming the landscape of health care.Mithrl is pioneering the world's first commercially available AI Co-Scientist, a revolutionary discovery engine that converts complex biological data into actionable insights in mere minutes. By simply asking questions in natural language, scientists receive immediate responses with comprehensive analyses, innovative targets, hypotheses, and patent-ready reports.Our impressive growth trajectory includes:12X year-over-year revenue growthTrusted by top-tier biotech firms and major pharmaceutical companies across three continentsFacilitating real breakthroughs from target discovery to patient outcomes.About the RoleWe are in search of an innovative Lead Bioinformatics Pipeline Engineer to architect and expand Mithrl’s multimodal scientific processing pipelines. You will be responsible for developing workflows that convert raw biological data into clean, reproducible outputs that fuel Mithrl’s AI Co-Scientist. Your work will encompass a range of modalities including microarray, imaging, spatial transcriptomics, genomics, epigenomics, flow cytometry, and more.This pivotal role lies at the core of our technical infrastructure. You will design Nextflow and nf-core style pipelines, implement modality-specific validation and quality control layers, and collaborate closely with our Tabular Data Team and Knowledge Curation Team to ensure seamless data harmonization, variable ID mapping, and schema alignment. Your contributions will enable scientists to pose inquiries and receive accurate, data-driven answers instantaneously.If you are passionate about constructing robust scientific workflows and tackling high-impact challenges, you will find your place here.
Full-time|$185K/yr - $235K/yr|On-site|San Francisco
About Stand Insurance Stand Insurance is rethinking how property risks are understood and managed. By combining advanced physics with artificial intelligence, the team models catastrophic risks at the asset level and automates underwriting and risk mitigation before losses happen. Instead of simply delivering insurance, Stand builds a scalable risk engine that aims to deliver real-world impact and stay in markets where others exit. Traditional property insurance often relies on outdated data and manual workflows, accepting damage as a given. Stand takes a different path: simulating real-world catastrophes for individual properties, turning those simulations into actionable steps, and automating operations around those insights. The result is a platform that can underwrite risks others avoid, while reducing operational friction. Role Overview: Machine Learning Engineer – Data Pipeline This role centers on building and maintaining the tools behind Stand’s data annotation pipeline. Areas of focus include computer vision, human-in-the-loop management, quality assurance, and economic optimization. The main goal: increase automation and lower cost-per-policy, while keeping quality high. Early on, work will involve hands-on management of the pipeline, quality checks, and close coordination with the annotation team. As experience grows, the focus will shift to developing advanced data science and machine learning systems, especially around quality instrumentation, automated QA, predictive labeling, and computer vision models. Over time, the role will evolve into shaping a systems-driven, automation-focused framework for the entire annotation lifecycle. Key Responsibilities Pipeline Operations and Reliability Monitor and maintain the daily health of the annotation pipeline Set up escalation protocols and frameworks for categorizing failures Lead the transition from manual to automated operations Quality Instrumentation Design validation systems that align with downstream model metrics Develop anomaly detection models for annotation workflows Automate tasks to cut down on manual QA effort Vendor and Annotator Performance Define and track performance metrics for vendors and annotators Location San Francisco
Full-time|$103.5K/yr - $196K/yr|Hybrid|San Francisco
About Our Organization:Welcome to Scribd Inc. (pronounced “scribbed”), where our passion lies in igniting human curiosity through storytelling and knowledge-sharing. We invite you to join our dynamic team as we work towards democratizing the exchange of ideas and empowering collective expertise with our innovative products: Everand, Scribd, Slideshare, and Fable.This job posting represents an established opportunity within our organization.At Scribd, we cultivate a culture where authenticity and boldness thrive. We value open discussions and commitment as we embrace the unexpected, empowering every employee to take initiative while keeping our customers at the forefront.We believe in a balanced approach to work structure, merging individual flexibility with community engagement. Our Scribd Flex program allows employees, in collaboration with their managers, to choose work styles that best suit their needs. This initiative emphasizes the importance of intentional in-person gatherings to foster collaboration and connection. Thus, occasional in-person attendance is a requirement for all Scribd employees, regardless of their remote status.What do we seek in our new teammates? We prioritize candidates who embody “GRIT” – a blend of passion and perseverance towards long-term goals. At Scribd, we encourage a GRIT-driven approach to work, where the ability to set and achieve Goals, deliver Results, contribute Innovative ideas, and positively impact the Team through collaboration is essential.About the Team:Our ML Data Engineering team is at the forefront of metadata extraction, enrichment, and content understanding across all Scribd products. We manage vast volumes of documents and images, ensuring high-quality metadata that enhances content discovery and builds trust among millions of users around the globe.Our systems function on a massive scale, incorporating diverse datasets like user-generated content, ebooks, audiobooks, and more. We operate at the convergence of machine learning, data engineering, and distributed systems, working closely with applied research and product teams to deploy scalable ML solutions.
Full-time|$103.5K/yr - $196K/yr|Hybrid|San Francisco
At Scribd, Inc., we aim to enhance human understanding. Our innovative products—Scribd®, Slideshare®, Everand™, and Fable—empower billions globally to not just access knowledge, but also to apply it and achieve expertise.Company CultureWe foster an environment where our employees can be authentic and courageous, engaging in constructive debates while embracing unexpected challenges. Every team member is encouraged to take initiative, prioritizing customer needs.We understand that optimal performance arises from a mix of personal flexibility and meaningful community interaction. Our Scribd Flex program allows team members to choose their working style and location, while still emphasizing in-person collaborations that enrich our culture. Attendance at occasional in-person events is necessary for all employees, regardless of their location.We seek team members who embody “GRIT”—a blend of passion and perseverance towards long-term objectives. This ethos informs our approach to setting and achieving Goals, delivering impactful Results, fostering Innovation, and enhancing our Team dynamics through collaboration.This posting represents an open position within our organization.About Our Team:Our ML Data Engineering team is responsible for metadata extraction, enrichment, and content comprehension across all Scribd offerings. We handle hundreds of millions of documents and billions of images, providing high-quality metadata to facilitate content discovery and trust for millions of users worldwide.Operating at a massive scale, our systems support diverse datasets, including user-generated content (UGC), ebooks, audiobooks, and more. We collaborate closely with applied research and product teams to deploy scalable machine learning and large language model (LLM)-powered solutions in production.Role Overview:We are looking for a Software Engineer II with robust backend development expertise and a keen interest in addressing complex data challenges at scale. You will design, build, and optimize distributed systems that extract, enrich, and process metadata for a variety of content. This role involves close collaboration with ML engineers, product managers, and cross-functional teams.
At worldlabs, we are pioneers in building foundational world models that possess the ability to perceive, generate, reason, and interact seamlessly with the 3D environment. Our mission is to unlock the full potential of AI through spatial intelligence, transforming perception into action and imagination into reality.We envision that spatial intelligence will revolutionize storytelling, creativity, design, simulation, and immersive experiences in both the virtual and physical realms.We are proud to bring together a world-class team driven by curiosity and passion, with diverse backgrounds in technology—from AI research to systems engineering and product design—creating a dynamic feedback loop between our innovative research and user-centric products.Role OverviewWe are in the process of developing a robust web application focused on 3D Gaussian Splat scene generation, editing, and publishing. We seek a Pipeline Engineer who will integrate cutting-edge research features, ensuring they are reliable, debuggable, and user-friendly.This role is characterized by a high level of ownership and is primarily backend-focused, bridging the gap between R&D and frontend. You will engage end-to-end with graphics/ML algorithms, backend services, and frontend UI, transforming proof-of-concepts into dependable features. We are looking for someone who thrives on making complex systems operate smoothly in production and continuously refines them based on both internal evaluations and user feedback.
About World Labs: At World Labs, we are pioneers in building foundational world models that can perceive, generate, reason, and engage with the 3D environment. Our mission is to unlock the full potential of artificial intelligence through spatial intelligence, transforming vision into action, perception into reasoning, and imagination into creativity. We believe that spatial intelligence will pave the way for new forms of storytelling, creativity, design, simulation, and immersive experiences across both virtual and physical realms. Our team is composed of exceptional talent united by a shared curiosity and passion for technology, ranging from AI research to systems engineering and product design. Together, we create a dynamic feedback loop between our cutting-edge research and innovative products that empower our users. Role Overview We are in search of a dedicated 3D Data Pipeline Engineer to design, build, and manage the critical systems that facilitate high-quality 3D data processing, synthetic data generation, and rendering across our suite of products. This hands-on role is ideal for someone enthusiastic about large-scale 3D data, system performance, and establishing reliable data pipelines to enhance our product features. In this position, you will collaborate closely with product engineers, 3D artists, and research scientists to develop efficient, robust, and scalable data pipeline capabilities while ensuring high data integrity and performance in our fast-paced startup environment.
About AlembicAlembic is at the forefront of transforming marketing strategies, demonstrating the actual ROI of marketing initiatives. Our cutting-edge Alembic Marketing Intelligence Platform employs advanced algorithms and AI models to address this longstanding challenge effectively. By joining our team, you'll contribute to the development of tools that deliver unparalleled insights into how marketing influences revenue, empowering a growing roster of Fortune 500 companies to make data-driven decisions with confidence.About the RoleIn your role as a Senior Data Engineer at Alembic, you will play a crucial role in our data platform. You will be responsible for creating scalable and dependable data pipelines, optimizing storage solutions, and facilitating both real-time and batch analytics. Collaborating closely with data scientists, software engineers, and product leaders, you will design and implement robust data architectures that propel our mission forward.Key ResponsibilitiesDesign, develop, and maintain scalable ETL pipelines that efficiently ingest, process, and transform extensive volumes of structured and unstructured data.Optimize data storage solutions utilizing modern data lakehouse architectures and industry best practices to enhance cost-effectiveness, performance, and reliability.Collaborate with data scientists and engineers to seamlessly integrate machine learning models and analytical workloads into production environments.Ensure the integrity, quality, and security of data by implementing monitoring, alerting, and governance best practices.Work with cloud-based data warehouses and distributed data processing frameworks to support our data initiatives.Continuously assess and implement innovative technologies to enhance data infrastructure and operational efficiency.What We’re Looking For10+ years of experience in data engineering, software engineering, or a related field.Strong proficiency in SQL and Python for data processing.Experience with contemporary data warehousing and lakehouse solutions (e.g., Iceberg or similar).Expertise in distributed systems and big data technologies (Apache Spark, Hadoop, Kafka, Flink).Hands-on experience with cloud platforms (AWS, GCP, Azure) and related data services.Deep understanding of data management and governance practices.
ABOUT UNTO LABSAt Unto Labs, we are a team of visionary engineers committed to pushing the boundaries of distributed systems and developing cutting-edge blockchain technology on standard hardware. Our focus lies in creating core system primitives and high-performance architectures that will shape the future of digital assets and markets.ABOUT THE ROLEWe are seeking a skilled Platform/DevOps Engineer to take ownership of the infrastructure utilized by our engineering team. You will be responsible for managing CI/CD pipelines, deployment environments, observability, and developer tools. Our infrastructure is rapidly evolving, encompassing distributed node architectures across multiple cloud providers, self-managed CI runners, and a variety of deployment targets. Your contributions will enhance the reliability, performance, and cost-effectiveness of our systems.This is a high-impact position within a small, elite team. Your efforts will empower every engineer to work quickly and minimize disruptions. You can expect to collaborate directly with systems engineers and product teams to build and fortify the platform layer for next-generation blockchain systems.RESPONSIBILITIESTake charge of optimizing our CI/CD pipelines to ensure reliability, cost efficiency, and effective spin-down strategies to manage compute costs.Consolidate and streamline deployment infrastructure across Cloudflare, Google Cloud Run, and legacy providers, minimizing redundancy and fragility.Implement and maintain observability through structured logging (OpenTelemetry/HyperDX), alerting (PagerDuty/Slack), and dashboards that monitor node health, transaction flow, and network performance.Facilitate smoother developer onboarding by creating reproducible local development setups, improving secrets management, and maintaining clear documentation.Support network operations, manage chain state, automate backups, and maintain high uptime standards.Assess and deploy infrastructure tooling, including Infrastructure as Code (IaC), container orchestration, and cloud cost monitoring.REQUIREMENTSMinimum of 3 years of experience in platform engineering, DevOps, SRE, or similar infrastructure roles, preferably in high-growth or startup environments.Proficient in CI/CD systems, including self-hosted runners and cost tracking.Extensive knowledge of Google Cloud Platform (GCP) (including Cloud Run, Compute Engine, GCS, IAM) or equivalent AWS services.Experienced with Docker/containerization and related technologies.
About the CompanyBraintrust is at the forefront of AI observability. By seamlessly integrating evaluations and observability in a unified workflow, Braintrust empowers developers with the insights necessary to comprehend AI behavior in production and provides tools to enhance it.Renowned teams from Notion, Stripe, Zapier, Vercel, and Ramp utilize Braintrust to analyze models, test prompts, and detect regressions, transforming production data into superior AI with every iteration.About the RoleWe are seeking a highly skilled and proactive data professional to take charge of our data infrastructure and pipeline. This individual will possess extensive technical expertise in data engineering and the initiative to address challenges throughout the entire data stack. If you excel in problem-solving, are eager to learn new technologies, and take action without waiting for approval, this position is perfect for you.What You'll DoDevelop and maintain robust data pipelines that drive our analytics and business operations.Oversee our Snowflake and dbt infrastructure—managing data warehouse architecture, optimizing performance, and ensuring clean, well-documented models.Establish and troubleshoot data connectors across diverse sources and systems.Provide rapid analytics and dashboards to address business inquiries.Facilitate collaboration and communication to remove obstacles through coding or teamwork.Experiment with new tools and technologies, even those you haven't previously encountered.What We're Looking ForProven track record in data engineering—your portfolio is as important as your years of experience.In-depth knowledge of Snowflake, dbt management, and various data connectivity solutions.Comfortable with setting up data connectors, generating quick analytics, and crafting dashboards across the data stack.Exhibit high agency—identify issues, propose solutions, and deliver results independently.Skilled in unblocking yourself and others through effective communication and collaboration or by diving into code.Willing to explore uncharted territories, tackling new challenges head-on.Adaptable to ambiguity and adept at navigating problems as they arise.Benefits IncludeMedical, dental, and vision insurance401k
Full-time|$209K/yr - $253K/yr|On-site|San Francisco, CA - US
At Crusoe, we are dedicated to accelerating the abundance of energy and intelligence. Our innovative solutions are designed to empower a world where ambitious AI creations thrive, seamlessly merging scale, speed, and sustainability.Join us in the AI revolution by leveraging sustainable technology at Crusoe. Here, you will spearhead significant innovations, make a real difference, and collaborate with a team that is leading the charge in responsible and transformative cloud infrastructure.About This Role:We seek a dynamic Engineering Manager to guide our Managed Services and Production Engineering team. In this role, you will balance your efforts between enhancing reliability and managing a team of Site Reliability Engineers (SREs) focused on Crusoe's AI-managed and service offerings. You will take responsibility for the production health of services provided to enterprise clients, including Managed Kubernetes, Managed Inference, and AutoClusters, while collaborating closely with embedded engineering teams to elevate operational excellence, automation, and customer satisfaction.What You'll Be Working On:Leading and developing a team of SREs within Crusoe's Managed AI and Managed Services sectors, establishing technical direction and nurturing a culture of ownership and continuous enhancement.Contributing as an individual contributor—reviewing code, creating tools, and promoting automation to minimize manual work and increase the reliability and scalability of managed services.Taking charge of SLA/SLO performance, incident response, and on-call health for managed services; conducting blameless post-mortems and driving systemic remediation efforts.Collaborating with product and platform engineering teams to influence infrastructure design, observability strategies, and operational readiness for both new and existing managed services.Establishing and monitoring reliability, performance, and operational maturity metrics across the team; converting data into prioritized roadmap investments.Acting as a technical escalation point for high-severity production incidents impacting enterprise customers, and coordinating with Cloud Support and Customer Success teams to ensure effective resolution and communication.What You'll Bring to the Team:Over 5 years of experience in software or infrastructure engineering, with a minimum of 1–2 years in an engineering management or technical lead position.Proven track record in leading teams and driving technical initiatives.
About Our TeamThe Applied team at OpenAI is dedicated to collaborating across research, engineering, product, and design disciplines to effectively deliver cutting-edge AI technology to both consumers and businesses. Our mission is to learn from real-world deployments, ensuring that the benefits of AI are distributed equitably while prioritizing responsible and safe use of this powerful technology. Here, safety takes precedence over unchecked growth.About the Role:We are on the lookout for a Senior Data Engineer to spearhead the development of our data pipelines and foundational data tables at OpenAI. These pipelines are essential for driving analyses, enhancing safety systems that inform business decisions, fostering product growth, and mitigating risks from malicious actors. If you have a passion for data and a desire to create impactful solutions, we want to hear from you! This role also offers the chance to work closely with the innovative researchers behind ChatGPT, assisting them in training new models for user delivery. As we continue to expand rapidly, we place immense value on data-driven insights, and your contributions will be crucial to our ongoing success. Join us in shaping the future of AI!Your Responsibilities Will Include:Designing, building, and maintaining our data pipelines, ensuring seamless integration of user event data into our data warehouse.Creating canonical datasets that track vital product metrics such as user growth, engagement, and revenue.Collaborating with diverse teams, including Infrastructure, Data Science, Product, Marketing, Finance, and Research, to understand their data needs and deliver effective solutions.Implementing robust, fault-tolerant systems for data ingestion and processing.Participating in architectural and engineering decisions related to data management, leveraging your extensive experience.Ensuring data security, integrity, and compliance with industry standards and company policies.You Will Excel in This Role If You:Possess 3+ years of experience as a data engineer and at least 8 years in software engineering (including data engineering).Have proficiency in at least one programming language relevant to Data Engineering, such as Python, Scala, or Java.Bring experience with distributed processing technologies and frameworks like Hadoop, Flink, and distributed storage systems (e.g., HDFS, S3).Are well-versed in ETL schedulers such as Airflow, Dagster, Prefect, or similar tools.
At AppsFlyer, we empower brands to make informed decisions with cutting-edge measurement, data analytics, deep linking, engagement, fraud protection, data clean rooms, and privacy-centric technologies. Our mission is to enhance customer privacy while delivering outstanding experiences, enabling thousands of creators and over 10,000 technology partners to foster deeper, more meaningful connections with their customers. We are looking for a dedicated Clay Consultant - GTM Engineer to transform and enhance our AI-driven pipeline and SDR infrastructure. This role is a 6-month contract, focusing on transitioning from manual processes to automated, scalable systems, thereby facilitating RevOps ownership. This position is a hands-on builder role, where your time will primarily be spent executing within tools and systems rather than developing strategic presentations.
Full-time|Remote|Denver, Colorado, United States; San Francisco, California, United States
Join Checkr as a Senior Data Engineer and play a pivotal role in shaping our data infrastructure and analytics capabilities. In this position, you will collaborate with cross-functional teams to design, develop, and maintain scalable data processing systems that empower our business decisions. If you are passionate about harnessing the power of data and thrive in a dynamic environment, this is the opportunity for you!
About UsSieve is a pioneering AI research lab dedicated solely to video data. We harness exabyte-scale video infrastructure and innovative video understanding techniques, along with a multitude of data sources, to create datasets that advance the field of video modeling. Given that video constitutes 80% of internet traffic, it serves as a vital medium that fuels creativity, communication, gaming, AR/VR, and robotics. Our mission is to tackle the most significant challenge in the development of these applications: acquiring high-quality training data.With a small yet highly skilled team of just 15 members, we have formed strategic partnerships with leading AI labs and achieved $XXM in revenue last quarter alone. Our Series A funding round last year was backed by prestigious firms, including Matrix Partners, Swift Ventures, Y Combinator, and AI Grant.About the RoleAs a Distributed Systems Engineer at Sieve, you will be responsible for designing and implementing systems that efficiently manage the compute, scheduling, and orchestration of complex machine learning and ETL pipelines. Your work will ensure these systems operate quickly, reliably, and cost-effectively while processing large volumes of video data.You will thrive in this role if you are passionate about optimizing system uptime, have experience with cloud technologies, and enjoy working with high-performance distributed systems involving thousands of GPUs. Additionally, you will play a key role in developing excellent internal tools and CI/CD pipelines to facilitate rapid iteration.
Hevo Data (www.hevodata.com) offers an intuitive, no-code data pipeline platform that empowers organizations to integrate data across various software systems for accelerated analytics.With a commitment to driving data insights, Hevo supports over 2,000 data-centric companies across diverse sectors, including Shopify, Gartner, Iceland AIR, Arhaus, Thoughtspot, and Postman. By streamlining complex data integration processes, Hevo enables data teams to concentrate on uncovering innovative insights and propelling their businesses forward.At Hevo, we are on a mission to develop technology that is easy to adopt and accessible for all, unlocking the true potential of data.Based in San Francisco, with additional offices in India, Hevo has experienced unprecedented growth since its founding, tripling its revenue and customer base in just two years.With $42 million in funding from esteemed investors like Sequoia India, Qualgro, and Chiratae Ventures, Hevo is embarking on an exciting journey of hyper-growth.Our team, known as Hevoites, comprises thoughtful and proactive problem solvers dedicated to making a positive impact in the lives of our customers and colleagues, as well as advancing their personal careers.If you are passionate about transforming the future of technology, consider joining us at Hevo!Product Video: https://www.youtube.com/watch?v=p0XGLDgvCo8What
Join us at datacurve as we innovate a gamified developer platform that empowers thousands of engineers to create high-fidelity datasets, advancing the frontiers of large language models (LLMs). In this pivotal role, you will oversee the entire technical lifecycle of our data pipelines—from collaborating with partner labs to establish new data formats, to delivering the essential tools, environments, documentation, and quality assurance processes that bring these formats to life at scale.Key ResponsibilitiesLead Projects End-to-End: Take ownership of projects from initial prototyping through to ongoing maintenance and iterative improvements based on user feedback.Oversee Developer Experience Pipelines: Develop and prototype tools for capturing new data formats, transitioning to a production workflow, and refining the developer experience.Champion Developer Experience: Produce clear and concise guidelines and documentation to empower our contributors and ensure the quality of project inputs.Quality Assurance & Governance: Establish and manage quality standards for your projects, which includes training content reviewers to ensure data consistency and accuracy. Implement automated checks, evaluation harnesses, and workflows to meet data quality benchmarks.Continuous Improvement: Monitor systems, troubleshoot issues, and enhance reliability, latency, and contributor success rates.Occasional ResponsibilitiesDefine Innovative Data Formats: Collaborate with frontier lab researchers to create specifications and design schemas, metadata, and versioning for new formats.Develop Tools and Environments: Deliver tools, sandboxes, command-line interfaces (CLIs), and instrumentation to streamline contribution processes.
Full-time|$164K/yr - $227K/yr|On-site|San Francisco, CA, USA
Role overview Chime’s Data Engineering team develops the systems that power data-driven decisions across the company. Senior Data Engineers play a key role in designing and implementing scalable data pipelines and frameworks, making sure analytics remain reliable and well-governed. This work supports teams across Chime as they build new capabilities and improve how data informs business choices. What you will do Build and maintain scalable data pipelines and frameworks to support analytics Create solutions that keep data accessible, accurate, and governed Design workflows for analytics and reporting used throughout the organization Help shape data engineering practices that can influence fintech standards Compensation and benefits The base salary for this Senior Data Engineer position ranges from $164,000 to $227,000. Full-time employees are also eligible for bonuses, equity options, and a comprehensive benefits package. Final salary depends on skills, qualifications, and experience.
Join the Founding Team as a Data Engineer!Location: San Francisco Bay AreaType: Full-TimeCompensation: Competitive salary + early-stage equityAt fabrion, supported by 8VC, we are on a mission to assemble an elite team dedicated to solving one of the most significant challenges in the industry’s infrastructure.We are creating a cutting-edge, AI-native platform that transforms enterprise data into actionable insights through semantic enrichment, intelligent agents, and regulated interoperability. Central to our architecture is the Data Fabric—an intelligent, governed layer that integrates fragmented and siloed data into a cohesive ontology, primed for model training, vector search, and insight-to-action workflows.We seek engineers who thrive on tackling complex data challenges at scale, including messy unstructured data, schema evolution, multi-source integrations, security frameworks, and AI-ready semantic enrichment. You will be instrumental in developing backend systems, data pipelines, connector frameworks, and graph-based knowledge models that drive agentic applications.If your experience includes working on streaming unstructured data pipelines, integrating with challenging legacy systems, or scaling knowledge graphs, this opportunity is tailor-made for you!
Aug 11, 2025
Sign in to browse more jobs
Create account — see all 8,088 results
Tailoring 0 resumes…
Tailoring 0 resumes…
We'll move completed jobs to Ready to Apply automatically.