About the job
Cerebras Systems is at the forefront of AI technology, manufacturing the world’s largest AI chip, which is 56 times larger than traditional GPUs. Our innovative wafer-scale architecture delivers the AI computing power comparable to dozens of GPUs on a single chip, while maintaining the programming simplicity of a single device. This groundbreaking approach enables Cerebras to achieve unmatched training and inference speeds, allowing machine learning professionals to seamlessly deploy large-scale ML applications without the complexities of managing multiple GPUs or TPUs.
Cerebras’ impressive client roster includes leading model labs, major global enterprises, and pioneering AI-driven startups. Recently, OpenAI announced a multi-year partnership with Cerebras, aimed at harnessing 750 megawatts of scale to revolutionize key workloads through ultra high-speed inference.
Leveraging our cutting-edge wafer-scale architecture, Cerebras Inference offers the fastest Generative AI inference solution globally, boasting speeds over 10 times faster than GPU-based hyperscale cloud inference services. This dramatic increase in speed is transforming the user experience for AI applications, facilitating real-time iteration and enhancing intelligence through advanced agentic computation.
The Role
Join our Embedded Software team to contribute to the critical software framework that empowers the Cerebras Wafer Scale technology. You will work on innovative projects that push the boundaries of AI and embedded systems development, collaborating with a talented group of engineers focused on delivering exceptional performance for our clients.

