About the job
Cerebras Systems is at the forefront of AI innovation, having developed the world's largest AI chip, which is 56 times greater in size than conventional GPUs. Our revolutionary wafer-scale architecture delivers the computational power of multiple GPUs on a single chip, simplifying programming to a single device experience. This unique approach enables Cerebras to provide unparalleled training and inference speeds, allowing machine learning professionals to seamlessly operate large-scale ML applications without the complexities of managing numerous GPUs or TPUs.
Our clientele includes leading model labs, global corporations, and pioneering AI-native startups. Recently, OpenAI formed a multi-year collaboration with Cerebras to harness 750 megawatts of capacity, revolutionizing critical workloads with ultra-fast inference capabilities.
Thanks to our innovative wafer-scale architecture, Cerebras Inference stands as the fastest Generative AI inference solution globally, boasting speeds over ten times faster than traditional GPU-based hyperscale cloud inference services. This significant enhancement in speed transforms user experiences with AI applications, facilitating real-time iterations and augmenting intelligence through additional agentic computation.
About The Role
In the capacity of a Senior Software Engineer within the ML Integration and Quality team, you will be instrumental in integrating and delivering all software and hardware components of the Cerebras AI platform. Your focus will be on software feature integration and quality assurance, including pre-deployment and production validation of Cerebras' training and inference solutions. You will advocate for superior testing practices, effective debugging methodologies, and exemplary cross-team communication to ensure the delivery of world-class products.

