About the job
At Cerebras Systems, we are revolutionizing AI computing by developing the world’s largest AI chip, which is 56 times larger than traditional GPUs. Our innovative wafer-scale architecture delivers the computational prowess of dozens of GPUs on a single chip, combining remarkable performance with the simplicity of programming a single device. This cutting-edge approach enables us to provide unparalleled training and inference speeds, allowing machine learning practitioners to seamlessly execute large-scale ML applications without the complexities of managing numerous GPUs or TPUs.
Cerebras proudly serves a diverse clientele that includes leading model labs, global enterprises, and innovative AI-native startups. Recently, OpenAI announced a multi-year partnership with Cerebras, committing to deploy 750 megawatts of scale to transform critical workloads through ultra high-speed inference.
Thanks to our groundbreaking wafer-scale architecture, Cerebras Inference represents the fastest Generative AI inference solution globally, achieving speeds over 10 times faster than GPU-based hyperscale cloud inference services. This significant enhancement in speed is reshaping the user experience of AI applications, facilitating real-time iterations, and augmenting intelligence through enhanced computational capabilities.
About the Role
As a Full Stack Engineer concentrating on Cerebras’ manufacturing test platform, you will be responsible for designing, developing, and maintaining a comprehensive testing software solution for all phases of manufacturing—from individual components to complete Cerebras systems. You will collaborate across various teams including hardware design, engineering, operations, and data analytics to create user interfaces and data processing frameworks that significantly enhance manufacturing efficiency, quality, and scalability.

