Qualifications
Required ExperienceDemonstrated expertise in architecting, building, and managing large-scale data systems (petabyte scale). Familiarity with both batch and real-time processing architectures. Proficiency with open-source Data Lakehouse technologies such as Apache Iceberg, PostgreSQL, Neo4j, and Apache Parquet. Experience with stream processing and data analytics tools including Apache Kafka, Spark, and Flink. Solid understanding and hands-on experience with data transformation solutions. Strong programming skills in Python. Excellent communication and documentation abilities. Experience with cloud service provider (CSP) data platforms is a plus. Familiarity with data lineage, quality, and governance tools is a plus.
About the job
Your Impact
Become a key member of our innovative team at Cylake Inc., where we are dedicated to designing cutting-edge cybersecurity solutions from the ground up. You will collaborate with industry experts to architect, develop, and launch impactful products that are set to revolutionize the cybersecurity landscape. This is an exciting opportunity to advance your career alongside a dynamic team committed to excellence.
Role Overview
As a Data Pipeline Engineer, you will be responsible for designing, constructing, and sustaining a robust, open-source data lakehouse architecture capable of supporting petabyte-scale analytics. Your role will involve creating end-to-end data pipelines that ensure seamless data ingestion, transformation, and consumption while maintaining high performance, reliability, and data integrity.
About Cylake Inc.
Cylake Inc. is at the forefront of cybersecurity innovation, bringing together a team of seasoned professionals with a track record of success. We are committed to creating a diverse and inclusive work environment where every employee can thrive and contribute to impactful projects.