Cerebras just built a big chip that could democratize AI

Chip startup Cerebras has developed a foot-wide piece of silicon, compared to average chips measured in millimeters, that makes training AI cheap and easy.

A wafer-sized chip made by Cerebras Systems

At the core of Cerebras’ pitch is a chip that is roughly the size of a dinner plate.

Photo: Cerebras Systems

Inside a conference room at a Silicon Valley data center last week, chip startup Cerebras Systems founder and CEO Andrew Feldman demonstrated how the company’s technology allows people to shift between deploying different versions of an AI natural language model in a matter of moments, a task that usually takes hours or days.

“So we’ve made it 15 keystrokes to move among these largest models that have ever been described on a single machine,” Feldman said.

This, to Feldman and Cerebras, represents a triumph worth noting. Cerebras claims the system that achieved this feat has also accomplished a world first: It can train an entire 20-billion-parameter model on a single nearly foot-wide superchip. Without its technology, the company said scaling an AI model from 1 billion parameters to 20 billion parameters might require users to add more server hardware and reconfigure racks inside of a data center.

Training a natural language AI model on one chip makes it considerably cheaper and delivers a performance boost that is an order of magnitude superior to Nvidia’s flagship graphics processor-based systems, Feldman said. The idea is to give researchers and organizations with tiny budgets — in the tens of thousands of dollars range — access to AI training tools that were previously only available to much larger organizations with lots of money.

“Models have grown really fast in this area. Language processing, and the challenges of delivering compute for these models, is enormous,” Feldman said. “We sort of have made this class of model practical, useful to a whole slice of the economy that couldn’t previously do interesting work.”

The AI models that Feldman is talking about are simply methods of organizing mathematical calculations by breaking them up into steps and then regulating the communication between the steps. The point is to train a model to begin to make accurate predictions, whether that’s the next piece of code that should be written, what constitutes spam and so on.

AI models are typically large to begin with, but those built around language tend to be even larger. For language models, context — as in more text, such as adding an author’s entire body of work to a model that began with a single book — is crucial, but that context can make them far, far more complex to operate. Market-leader Nvidia estimates that AI tasks have spurred a 25-fold increase in the need for processing power every two years.

This exponential increase has led companies like Cerebras and others to chase AI as a potential market. For years, hardware investments were seen as bad bets among venture capitalists who were only willing to fund a few promising ideas. But as it became clear that AI as a class of computation would open the door for fresh ideas beyond the general purpose processors made by the likes of Intel and Nvidia, a new class of startups was born.

Cerebras, which is Latin for “mind,” is one of those startups. Founded in 2015, Feldman and his team, which includes a number of AMD veterans in key technology roles, have raised roughly $735 million — including funding from the CIA venture arm In-Q-Tel, the CEO said — at a $4.1 billion valuation.

Chips on the plate

At the core of Cerebras’ pitch is a chip that is roughly the size of a dinner plate, or an entire foot-wide silicon wafer, called the Wafer Scale Engine.

The idea of a wafer-size chip like the one that powers Cerebras’ systems isn’t a novel concept; similar ideas have been floating around for decades. A failed bid by Trilogy Systems in the early 1980s that raised roughly $750 million in today’s dollars is one notable attempt at a superchip, and IBM and others have studied the idea but never produced a product.

But together with TSMC, Cerebras has settled on a design that could be fabricated into a functioning wafer-size chip. In some ways, Cerebras is almost two startups stuck together: It’s interested in tackling the growing challenge of AI compute, but it has also achieved the technological feat of producing a useful chip the size of a wafer.

the cerebras cs-2 server system in a data center A Cerebras CS-2 system running inside a data center.Photo: Max A. Cherney/Protocol

The current generation of what Cerebras calls the WSE-2 can offer considerable performance improvements over stringing together multiple graphics chips to achieve the computational horsepower to train some of the largest AI models, according to Feldman.

“So it's unusual for a startup to have deep fab expertise, [but] we have profound expertise,” Feldman said. “And we had an idea of how they could, within their permitted flexibility in their flow, fit our innovation.”

The advantage of building a chip of that size is that it allows Cerebras to duplicate the performance of dozens of other server chips — roughly 80 graphics processors, for some large AI models — and squishes them onto a single piece of silicon. Doing so makes them considerably faster, because, in part, data can move faster across a single chip than across a network of dozens of chips.

"[Our] machine is built for one type of work,” Feldman said. “If you want to take the kids to soccer practice, no matter how shitty they are to drive, the minivan is the perfect car. But if you've got your minivan and you try and move two-by-fours and 50-pound sacks of concrete, you realize what a terrible machine it is for that job. [Our chip] is a machine for AI.”

This story was updated to correct the amount of money raised by Trilogy Systems.


Niantic’s future hinges on mapping the metaverse

The maker of Pokémon Go is hoping the metaverse will deliver its next big break.

Niantic's new standalone messaging and social app, Campfire, is a way to get players organizing and meeting up in the real world. It launches today for select Pokémon Go players.

Image: Niantic

Pokémon Go sent Niantic to the moon. But now the San Francisco-based augmented reality developer has returned to earth, and it’s been trying to chart its way back to the stars ever since. The company yesterday announced layoffs of about 8% of its workforce (about 85 to 90 people) and canceled four projects, Bloomberg reported, signaling another disappointment for the studio that still generates about $1 billion in revenue per year from Pokémon Go.

Finding its next big hit has been Niantic’s priority for years, and the company has been coming up short. For much of the past year or so, Niantic has turned its attention to the metaverse, with hopes that its location-based mobile games, AR tech and company philosophy around fostering physical connection and outdoor exploration can help it build what it now calls the “real world metaverse.”

Keep Reading Show less
Nick Statt

Nick Statt is Protocol's video game reporter. Prior to joining Protocol, he was news editor at The Verge covering the gaming industry, mobile apps and antitrust out of San Francisco, in addition to managing coverage of Silicon Valley tech giants and startups. He now resides in Rochester, New York, home of the garbage plate and, completely coincidentally, the World Video Game Hall of Fame. He can be reached at nstatt@protocol.com.

Every day, millions of us press the “order” button on our favorite coffee store's mobile application: Our chosen brew will be on the counter when we arrive. It’s a personalized, seamless experience that we have all come to expect. What we don’t know is what’s happening behind the scenes. The mobile application is sourcing data from a database that stores information about each customer and what their favorite coffee drinks are. It is also leveraging event-streaming data in real time to ensure the ingredients for your personal coffee are in supply at your local store.

Applications like this power our daily lives, and if they can’t access massive amounts of data stored in a database as well as stream data “in motion” instantaneously, you — and millions of customers — won’t have these in-the-moment experiences.

Keep Reading Show less
Jennifer Goforth Gregory
Jennifer Goforth Gregory has worked in the B2B technology industry for over 20 years. As a freelance writer she writes for top technology brands, including IBM, HPE, Adobe, AT&T, Verizon, Epson, Oracle, Intel and Square. She specializes in a wide range of technology, such as AI, IoT, cloud, cybersecurity, and CX. Jennifer also wrote a bestselling book The Freelance Content Marketing Writer to help other writers launch a high earning freelance business.

Supreme Court takes a sledgehammer to greenhouse gas regulations

The court ruled 6-3 that the EPA cannot use the Clean Air Act to regulate power plant greenhouse gas emissions. That leaves a patchwork of policies from states, utilities and, increasingly, tech companies to pick up the slack.

The Supreme Court struck a major blow to the federal government's ability to regulate greenhouse gases.

Eric Lee/Bloomberg via Getty Images

Striking down the right to abortion may be the Supreme Court's highest-profile decision this term. But on Thursday, the court handed down an equally massive verdict on the federal government's ability to regulate greenhouse gas emissions. In the case of West Virginia v. EPA, the court decided that the agency has no ability to regulate greenhouse gas pollution under the Clean Air Act. Weakening the federal government's powers leaves a patchwork of states, utilities and, increasingly, tech companies to pick up the slack in reducing carbon pollution.

Keep Reading Show less
Brian Kahn

Brian ( @blkahn) is Protocol's climate editor. Previously, he was the managing editor and founding senior writer at Earther, Gizmodo's climate site, where he covered everything from the weather to Big Oil's influence on politics. He also reported for Climate Central and the Wall Street Journal. In the even more distant past, he led sleigh rides to visit a herd of 7,000 elk and boat tours on the deepest lake in the U.S.


Can crypto regulate itself? The Lummis-Gillibrand bill hopes so.

Creating the equivalent of the stock markets’ FINRA for crypto is the ideal, but experts doubt that it will be easy.

The idea of creating a government-sanctioned private regulatory association has been drawing more attention in the debate over how to rein in a fast-growing industry whose technological quirks have baffled policymakers.

Illustration: Christopher T. Fong/Protocol

Regulating crypto is complicated. That’s why Sens. Cynthia Lummis and Kirsten Gillibrand want to explore the creation of a private sector group to help federal regulators do their job.

The bipartisan bill introduced by Lummis and Gillibrand would require the CFTC and the SEC to work with the crypto industry to look into setting up a self-regulatory organization to “facilitate innovative, efficient and orderly markets for digital assets.”

Keep Reading Show less
Benjamin Pimentel

Benjamin Pimentel ( @benpimentel) covers crypto and fintech from San Francisco. He has reported on many of the biggest tech stories over the past 20 years for the San Francisco Chronicle, Dow Jones MarketWatch and Business Insider, from the dot-com crash, the rise of cloud computing, social networking and AI to the impact of the Great Recession and the COVID crisis on Silicon Valley and beyond. He can be reached at bpimentel@protocol.com or via Google Voice at (925) 307-9342.


Alperovitch: Cybersecurity defenders can’t be on high alert every day

With the continued threat of Russian cyber escalation, cybersecurity and geopolitics expert Dmitri Alperovitch says it’s not ideal for the U.S. to oscillate between moments of high alert and lesser states of cyber readiness.

Dmitri Alperovitch (the co-founder and former CTO of CrowdStrike) speaks at RSA Conference 2022.

Photo: RSA Conference

When it comes to cybersecurity vigilance, Dmitri Alperovitch wants to see more focus on resiliency of IT systems — and less on doing "surges" around particular dates or events.

For instance, whatever Russia is doing at the moment.

Keep Reading Show less
Kyle Alspach

Kyle Alspach ( @KyleAlspach) is a senior reporter at Protocol, focused on cybersecurity. He has covered the tech industry since 2010 for outlets including VentureBeat, CRN and the Boston Globe. He lives in Portland, Oregon, and can be reached at kalspach@protocol.com.

Latest Stories