About the job
Cerebras Systems is at the forefront of AI technology, developing the world’s largest AI chip, which is 56 times larger than traditional GPUs. Our innovative wafer-scale architecture delivers the computational power of numerous GPUs within a single chip, coupled with the programming simplicity of a unified device. This groundbreaking approach enables us to achieve unparalleled training and inference speeds, allowing machine learning practitioners to effortlessly execute large-scale ML applications without the complexity of managing multiple GPUs or TPUs.
Our clientele comprises leading model laboratories, global corporations, and pioneering AI-centric startups. Recently, we forged a multi-year partnership with OpenAI to deploy 750 megawatts of scalable technology, revolutionizing critical workloads with ultra-high-speed inference.
Thanks to our revolutionary wafer-scale architecture, Cerebras Inference stands as the fastest Generative AI inference solution globally, achieving speeds over ten times that of GPU-based hyperscale cloud inference services. This significant increase in processing speed is transforming the user experience in AI applications, facilitating real-time iterations and enhancing intelligence through additional agentic computation.
About The Role
At Cerebras, we’re redefining the future of AI compute. We are seeking a Senior Technical Program Manager to spearhead complex, cross-functional programs across our AI training and inference platforms.

