← All jobs · Cerebras Systems

AI Engineering Intern - Growth Team

Cerebras Systems ·
86
AI-Agency
B82 U92
📍 Sunnyvale, US 🛠 AI tools welcome at work Intern Internship
PythonClaude CodeMCPRAGLLM APIsvector databases
TL;DR

AI Engineering Intern at Cerebras Systems building agentic workflows and internal developer tools for chip design, model bringup, and cloud platform teams. 12-week paid internship in Sunnyvale or Toronto using Claude Code, MCP, and RAG systems.

Apply at Cerebras Systems →
share:
you'll be redirected to the company's career page

Job description

Cerebras Systems builds the world's largest AI chip, 56 times larger than GPUs. Our novel wafer-scale architecture provides the AI compute power of dozens of GPUs on a single chip, with the programming simplicity of a single device. This approach allows Cerebras to deliver industry-leading training and inference speeds and empowers machine learning users to effortlessly run large-scale ML applications, without the hassle of managing hundreds of GPUs or TPUs.  

Cerebras' current customers include top model labs, global enterprises, and cutting-edge AI-native startups. OpenAI recently announced a multi-year partnership with Cerebras, to deploy 750 megawatts of scale, transforming key workloads with ultra high-speed inference. 

Thanks to the groundbreaking wafer-scale architecture, Cerebras Inference offers the fastest Generative AI inference solution in the world, over 10 times faster than GPU-based hyperscale cloud inference services. This order of magnitude increase in speed is transforming the user experience of AI applications, unlocking real-time iteration and increasing intelligence via additional agentic computation.

About the Team

The Growth Team drives AI adoption across Cerebras. We are a multi-disciplinary team that owns product, engineering and marketing responsibilities.

We build agentic workflows, internal knowledge systems, and developer infrastructure that help our engineering orgs – kernel, design verification, and cloud platform – ship faster. Our stack includes Claude Code, MCP (Model Context Protocol), RAG pipelines, and multi-agent architectures, and we work directly with the teams building Cerebras’s chips and inference platform.

This summer, you’ll embed with engineers across the company to build AI tooling that accelerate real hardware and software development workflows. Your work won’t sit on a shelf – you’ll ship internal tools that our engineers actually use.

About the Role

We’re looking for an AI Engineer Intern to join the Growth Team for Summer 2026. You’ll own end-to-end workstreams: scoping problems with engineering teams, building agentic systems, iterating based on user feedback, and shipping a working internal tool by the end of the internship.

This is a 12-week, paid, in-person internship based in our Sunnyvale, CA or Toronto, ON office, running June through August 2026.

What You’ll Work On

1. AI Agents for Design Verification & ASICs

Work directly with the design verification and ASICs teams to build AI agents that speed up chip development. This could include automated test generation, debug triage, or verification workflow acceleration. You’ll learn how silicon gets shipped and build tooling that compresses the iteration cycle.

2. AI Agents for Kernel & Model Bringup

Build agentic workflows that accelerate how we bring up new models on Cerebras hardware. This means working with the kernel team to identify bottlenecks in the bringup process and building AI-powered tools – using Claude Code, MCP integrations, and RAG systems – that help engineers move faster from first compile to production readiness.

3. AI Agents for Cloud Platform & SRE

Partner with the cloud platform team to build AI agents that reduce incident response time and speed up SRE workflows. Think automated log analysis, intelligent runbook execution, and agentic debug loops that help on-call engineers resolve issues faster.

Capstone: Ship an Internal Tool

By the end of the internship, you’ll have built and shipped at least one internal tool – whether that’s a RAG-powered knowledge base, an MCP integration, or a multi-agent system – that engineering teams at Cerebras are actively using. You’ll present your work and its impact to engineering leadership.

What We’re Looking For

Bonus:

What You’ll Get

Cerebras Systems is committed to creating an equal and diverse environment and is proud to be an equal opportunity employer. We celebrate different backgrounds, perspectives, and skills. We believe inclusive teams build better products and companies. We try every day to build a work environment that empowers people to do their best work through continuous learning, growth and support of those around them.

Why Join Cerebras

People who are serious about software make their own hardware. At Cerebras we have built a breakthrough architecture that is unlocking new opportunities for the AI industry. With dozens of model releases and rapid growth, we’ve reached an inflection  point in our business. Members of our team tell us there are five main reasons they joined Cerebras:

  1. Build a breakthrough AI platform beyond the constraints of the GPU.
  2. Publish and open source their cutting-edge AI research.
  3. Work on one of the fastest AI supercomputers in the world.
  4. Enjoy job stability with startup vitality.
  5. Our simple, non-corporate work culture that respects individual beliefs.

Read our blog: Five Reasons to Join Cerebras in 2026.

Apply today and become part of the forefront of groundbreaking advancements in AI!


Cerebras Systems is committed to creating an equal and diverse environment and is proud to be an equal opportunity employer. We celebrate different backgrounds, perspectives, and skills. We believe inclusive teams build better products and companies. We try every day to build a work environment that empowers people to do their best work through continuous learning, growth and support of those around them.


This website or its third-party tools process personal data. For more details, click here to review our CCPA disclosure notice.

Apply at Cerebras Systems →

More open roles at Cerebras Systems

Cerebras Systems · 🔄 synced 1h ago
Principal ML Investigator
📍 Sunnyvale, US 🛠 AI tools welcome at work · Principal
Principal ML Investigator at Cerebras Systems building ML research and advanced development on the company's wafer-scale AI chip platform. Focus on post-training, dataset optimization, LLM pretraining, sparsity, and domain-specific agents.
PyTorchJAXTensorFlowTritondistributed training frameworks
89
AI-core
Cerebras Systems · 🔄 synced 1h ago
Advanced Technology: AI/ML Research Scientist
📍 Sunnyvale, US · Senior
AI/ML Research Scientist at Cerebras Systems designing and developing AI models and training methodologies on wafer-scale hardware. Work spans optimization theory, model architecture, and computational science to explore new algorithmic possibilities enabled by novel hardware architecture.
PythonPyTorchC
87
AI-core
Cerebras Systems · 🔄 synced 1h ago
Applied AI/ML Scientist
📍 AE · Senior
Applied AI/ML Scientist at Cerebras Systems developing and customizing large language models on the company's wafer-scale AI chip. Responsibilities include training custom models, fine-tuning LLMs, designing agentic systems, and serving as technical expert for customer AI initiatives.
PythonPyTorchTransformersLLMsRLHFDPO
87
AI-core
Cerebras Systems · 🔄 synced 1h ago
LLM Inference Performance & Evals Engineer
📍 Toronto, CA 🛠 AI tools welcome at work · Mid
LLM Inference Performance & Evals Engineer at Cerebras Systems building performance-eval pipelines and benchmarking cutting-edge model optimizations on wafer-scale hardware. Prototype architectural tweaks, develop agent-driven automation for experiments, and collaborate across compiler, runtime, and silicon teams.
PythonTritonLLVMMLIRFlash-AttentionTransformers
83
AI-core
Cerebras Systems · 🔄 synced 1h ago
Advanced Technology: R&D Engineer - AI/ML, HPC
📍 Sunnyvale, US · Senior
R&D Engineer at Cerebras Systems designing and optimizing AI/ML and scientific computing workloads on wafer-scale hardware. Work spans algorithm design, compiler co-optimization, and hardware architecture to advance state-of-the-art performance benchmarks.
CPythonPyTorchCUDAHPCnumerical methods
79
AI-core