CS

AI Models, Product Manager

Verified
Cerebras Systems
Posted 2 weeks ago
Posted 15 April 2026
1 views
full-time

About the Role

<div class="content-intro"><p><span data-contrast="none">Cerebras Systems builds the world's largest AI chip, 56 times larger than GPUs. Our novel wafer-scale architecture provides the AI compute power of dozens of GPUs on a single chip, with the programming simplicity of a single device. This approach allows Cerebras to deliver industry-leading training and inference speeds and empowers machine learning users to effortlessly run large-scale ML applications, without the hassle of managing hundreds of GPUs or TPUs.&nbsp;</span><span data-ccp-props="{&quot;134233117&quot;:false,&quot;134233118&quot;:false,&quot;201341983&quot;:0,&quot;335559685&quot;:0,&quot;335559737&quot;:240,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:279}">&nbsp;</span></p> <p>Cerebras' current customers include top model labs, global enterprises, and cutting-edge AI-native startups.&nbsp;<a href="https://openai.com/index/cerebras-partnership/&quot;&gt;OpenAI recently announced a multi-year partnership with Cerebras</a>, to deploy 750 megawatts of scale, transforming key workloads with ultra high-speed inference.&nbsp;</p> <p>Thanks to the groundbreaking wafer-scale architecture, Cerebras Inference offers the fastest Generative AI inference solution in the world, over 10 times faster than GPU-based hyperscale cloud inference services. This order of magnitude increase in speed is transforming the user experience of AI applications, unlocking real-time iteration and increasing intelligence via additional agentic computation.</p></div><h4 id="Own-the-Future-of-AI-Inference" data-local-id="27e2ad7d-fed5-4399-8636-9d7785e16b0a" data-renderer-start-pos="568">Own the Future of AI Inference</h4> <p data-renderer-start-pos="600" data-local-id="d08d13f3-579c-4c92-a359-e166de311ab9">Cerebras powers the world's fastest AI inference. As the Product Manager for AI Models, you'll lead the strategic model portfolio that defines our product — deciding which models ship, how they perform, and how the world discovers them.</p> <p data-renderer-start-pos="838" data-local-id="4c18318d-cea4-42d1-83d7-1ecaf41c2c13">You'll partner directly with leading AI labs, drive launches that shape the industry, and ensure every model on our platform delivers exceptional quality at unprecedented speed.</p> <h4 id="What-You'll-Own" data-local-id="eeeca7a0-9050-41be-8319-446a655c863f" data-renderer-start-pos="1017">What You'll Own</h4> <h5 data-renderer-start-pos="1034" data-local-id="17edc850-f094-4f73-bac4-dfeb01e88274"><strong data-renderer-mark="true">Strategic Model Portfolio</strong></h5> <ul> <li data-renderer-start-pos="1063" data-local-id="d0add9b4-eed3-4ebd-b8ed-a5ec55d2771d">Own the models roadmap: decide which frontier and open-source models we support based on market demand, research trends, and strategic fit</li> <li>Establish partnerships with top model labs, for day0 launches</li> <li data-renderer-start-pos="1337" data-local-id="95183413-f570-4c16-aa98-01fe08d20b3e">Build relationships with open-source maintainers to accelerate community model adoption</li> </ul> <h5 data-renderer-start-pos="1428" data-local-id="e826256c-0d1e-4dca-856e-c215a14811f1"><strong data-renderer-mark="true">Product Quality &amp; Customer Success</strong></h5> <ul> <li data-renderer-start-pos="1466" data-local-id="588f966a-b91b-47e1-ad23-d6dd1ccc7aff">Define and enforce quality standards across our model catalog through systematic evaluation frameworks</li> <li data-renderer-start-pos="1572" data-local-id="4eac9d2c-357c-4600-b3a5-f2010e31a38b">Design benchmarks and evaluations that prove our models deliver production-grade performance</li> <li data-renderer-start-pos="1668" data-local-id="8499755a-f89b-4bdd-837c-b4cfe9529509">Own the feedback loop: gather customer insights, identify model weaknesses, and drive improvements with engineering</li> <li data-renderer-start-pos="1787" data-local-id="6d6c027f-6fa9-4566-bcba-d211a4c7e48c">Enable strategic customers to integrate our inference into their products—removing blockers and optimizing for their specific use cases</li> </ul> <h5 data-renderer-start-pos="1926" data-local-id="1f173345-61b0-4a9d-a1a7-a2f25777b965"><strong data-renderer-mark="true">Go-to-Market Excellence</strong></h5> <ul> <li data-renderer-start-pos="1953" data-local-id="97f8725a-d295-448a-933f-887961129253">Lead high-impact model launches that generate buzz and adoption</li> <li data-renderer-start-pos="2020" data-local-id="590f3a84-e495-446f-a067-7f561fa6bda1">Create compelling product marketing: demos, benchmarks, tutorials, and documentation that showcase what's possible on Cerebras</li> <li data-renderer-start-pos="2150" data-local-id="5959aee9-0d04-4141-a689-60a5627721b9">Craft technical content that resonates with developers and decision-makers alike</li> </ul> <h5 data-renderer-start-pos="2234" data-local-id="1380e20b-942a-4943-bf94-7ba94ad16ef7"><strong data-renderer-mark="true">Technical Decision-Making</strong></h5> <ul> <li data-renderer-start-pos="2263" data-local-id="8f72f63f-9009-4b7d-bfd7-85959b9f9356">Select and prioritize performance optimizations (quantization, speculative decoding, etc.) based on customer needs and hardware capabilities</li> <li data-renderer-start-pos="2407" data-local-id="23c1369d-9596-43a5-9dcc-917d707b4995">Collaborate with optimization engineers to implement techniques that maximize our speed advantage</li> <li data-renderer-start-pos="2508" data-local-id="9941a635-03b0-4e4e-9e0c-58166283ab6e">Balance tradeoffs between quality, latency, throughput, and cost</li> </ul> <h5 data-renderer-start-pos="2576" data-local-id="a14e7356-3ad2-41ff-a1f2-97ebd6ff9099"><strong data-renderer-mark="true">Cross-Functional Leadership</strong></h5> <ul> <li data-renderer-start-pos="2607" data-local-id="3d5647fe-0a96-4a21-954b-83f67f9d1beb">Orchestrate launches across model enablement, optimization engineering, deployment, sales, and marketing</li> <li data-renderer-start-pos="2715" data-local-id="a04def27-419a-4888-95d2-03689bd72e43">Drive alignment in a fast-moving environment where priorities shift based on model releases and customer needs</li> <li data-renderer-start-pos="2829" data-local-id="f9d57696-aa75-4225-87b8-326aaf2197b8">Be the voice of the customer to engineering and the voice of product to customers</li> </ul> <h4><span data-contrast="none"><span data-ccp-parastyle="heading 3">Skills &amp; </span><span data-ccp-parastyle="heading 3">Q</span><span data-ccp-parastyle="heading 3">ualifications </span></span><span data-ccp-props="{&quot;134245418&quot;:true,&quot;134245529&quot;:true,&quot;335559738&quot;:160,&quot;335559739&quot;:80}">&nbsp;</span></h4> <p data-renderer-start-pos="1822" data-local-id="624bebb2-0271-4f7d-a98c-5fc9affc8d4f"><strong data-renderer-mark="true">What we need to see: </strong></p> <ol> <li data-renderer-start-pos="1847" data-local-id="45617d6f-ef29-41e9-80f5-d074a04e6842">5+ years of experience as a product manager, currently at or above the level of Senior PM.</li> <li data-renderer-start-pos="1941" data-local-id="d51fcb21-99e4-4b45-9f50-7ceffaa4eafe">5+ years of total technical work experience (e.g. SWE, ML researcher, solution engineer).</li> <li data-renderer-start-pos="2034" data-local-id="875a46b2-7896-4226-aa1e-b6a37009e9a4">Ability to thrive in a fast-paced, dynamic environment. With an entrepreneurial sense of ownership and ability to lead projects.</li> <li data-renderer-start-pos="2166" data-local-id="175f896f-2ef2-47ba-bc51-b0da641da0f4">Knowledge and passion for the worlds of open-source models and generative AI research.</li> <li data-renderer-start-pos="2256" data-local-id="77b72ccf-a333-413e-8b6b-786125c80757">Knowledge of the community model ecosystem, including: PyTorch, Hugging Face, vLLM, and SGLang.</li> <li data-renderer-start-pos="2355" data-local-id="550d7bd6-45a8-4351-a03d-1ba918701868">Highly motivated, independent, organized, and an effective communicator.</li> <li data-renderer-start-pos="2431" data-local-id="83403608-1303-43b1-9930-93766de89439">Comfortable using Python with the chat completions API, for basic model testing.</li> </ol> <p><strong><span data-contrast="none">Preferred requirements</span></strong><span data-contrast="none"> </span><span data-ccp-props=&quo

Related Searches

Explore more opportunities matching this role's title, location, and skills.

Job Title PagesLocation PagesCompany PagesSkill Pages

Ready to apply?

Click below to apply directly on Cerebras Systems's careers page.

Get the top 10 hyper-growth roles delivered to your inbox every Tuesday.