companyCerebras Systems logo

Inference Frontend Engineer

Cerebras SystemsSunnyvale, CA
On-site Full-time

Clicking Apply Now takes you to AutoApply where you can tailor your resume and apply.


Unlock Your Potential

Generate Job-Optimized Resume

One Click And Our AI Optimizes Your Resume to Match The Job Description.

Is Your Resume Optimized For This Role?

Find Out If You're Highlighting The Right Skills And Fix What's Missing

Experience Level

Entry Level

Qualifications

Key ResponsibilitiesCollaborate with a team of elite engineers to tackle real-world challenges across the software stack. Design, implement, and test software solutions that significantly enhance system performance and user experience. Engage in learning and contributing across multiple layers of a fully integrated AI-accelerated system. Acquire practical experience with advanced hardware, compilers, distributed systems, and machine learning frameworks. Required QualificationsA recent graduate or a current student in a university program pursuing a degree in Computer Science, Computer Engineering, or a related field (graduation expected in 2026). This is a new graduate position. Demonstrated strong problem-solving abilities along with excellent communication skills. Proficiency in one or more programming languages; experience with C++ is advantageous.

About the job

Cerebras Systems is revolutionizing the AI landscape with the world's largest AI chip, which is 56 times more extensive than traditional GPUs. Our innovative wafer-scale architecture enables us to deliver the computational power of dozens of GPUs on a single chip, while offering the ease of programming like a single device. This groundbreaking approach empowers Cerebras to achieve unparalleled training and inference speeds, allowing machine learning practitioners to run large-scale ML applications effortlessly without the complexities of managing numerous GPUs or TPUs.

Cerebras serves a diverse clientele that includes leading model laboratories, global corporations, and pioneering AI-focused startups. Recently, OpenAI announced a multi-year collaboration with Cerebras to harness 750 megawatts of scale, significantly enhancing key workloads through ultra-fast inference capabilities.

With our cutting-edge wafer-scale architecture, Cerebras Inference provides the fastest Generative AI inference solution globally, exceeding the speed of GPU-based hyperscale cloud inference services by over ten times. This extraordinary speed transformation is reshaping the user experience of AI applications, facilitating real-time iterations and boosting intelligence through enhanced agentic computation.

About Cerebras Systems

Cerebras Systems is at the forefront of AI technology, recognized for producing the largest AI chip in the world, dramatically enhancing computational capabilities and simplifying the programming landscape for machine learning applications.

Similar jobs

Tailoring 0 resumes

We'll move completed jobs to Ready to Apply automatically.