About the job
Cerebras Systems is at the forefront of AI technology, creating the world's largest AI chip, which boasts a size 56 times larger than conventional GPUs. Our innovative wafer-scale architecture combines the computational power of numerous GPUs within a single chip, simplifying programming and maximizing efficiency. This unique approach enables us to deliver unparalleled training and inference speeds, empowering machine learning professionals to execute large-scale ML applications effortlessly, eliminating the need to manage multiple GPUs or TPUs.
Among our esteemed clientele are leading model labs, multinational corporations, and pioneering AI-native startups. Notably, OpenAI has recently formed a multi-year partnership with Cerebras, harnessing our technology to deploy 750 megawatts of power that revolutionize critical workloads through ultra-fast inference.
Our cutting-edge wafer-scale architecture places Cerebras Inference as the fastest Generative AI inference solution globally, achieving speeds more than 10 times faster than GPU-based hyperscale cloud services. This significant enhancement is reshaping the user experience in AI applications, facilitating real-time iterations and enriching intelligence through advanced computational capabilities.
Shape the Future of AI Inference
As the AI Models Product Manager, you will spearhead the strategic development of our model portfolio. This pivotal role involves determining which models are launched, assessing their performance, and ensuring they are easily discoverable.
Collaborating directly with leading AI laboratories, you will drive product launches that shape the industry and guarantee that every model on our platform achieves outstanding quality at unmatched speed.

