1/ Cerebras just built a chip with 50x the transistor count, 1,000x the memory and 10,000x the bandwidth of Nvidia’s flagship GPU. One such 'chip' could replace an entire rack of Nvidia GPUs.
What the heck is going on?
2/ It’s no coincidence that the fastest AI chip today—Nvidia’s V100—is also the largest @ 815mm^2. More area = more cores and memory.
But why not make the chip even bigger? Two reasons..
3/ First, the ASML lithography machines have a reticle size of 858mm^2. V100 is basically at the limit of what today's manufacturing allows.
4/ Second, defect rate. The larger the chip, the greater the chance for chip defects. As chip size grows, the defect rate -> 100%, yield -> 0%. It’s very hard to make large chips and still have enough working ones to sell.
5/ So what did Cerebras do? They made the largest chip ever—a chip the size of an entire wafer.
If this were pizza, Nvidia/Intel/AMD make 30cm pies and then cut them up and sell tiny 2cm^2 slices. Cerebras sells a giant uncut 21cm^2 slab.
6/ How did Cerebras get around the reticle limit and defect rate?
Their chips still use a conventional photomask and expose the wafer sections at a time. But they worked with TSMC to lay additional wires to wire all the chips together so they can work as a whole.
7/ Cerebras deals with defects through build-in redundancy. The chip has extra cores and I/O lanes built-in. If one core is a dud, data moves around it just as traffic is re-routed around a closed city block. According to Cerebras, only 1.5% of the hw is for redundancy.
8/ Why is one giant chip better than a bunch of tiny ones?
Because you can fit the whole neural network on one chip rather than dice it up and spread it across dozens of GPUs. It is easier to program, way faster, and far more power efficient.
Going to attempt this absurdly technical recipe from the Contra cookbook tonight and live tweet the process. Let’s begin!
First, make bay leaf oil. Dried bay leaf is notorious for having little flavor. But fresh bay is very good. The oil takes on a piney/eucalyptus note. Gorgeous color too.
I’m using skate wing to make the sauce. It needs to be rich and milky so I’m adopting a Chinese technique - fry and rapid boil to emulsify the fat. It’s still a bit thin so I might reduce more.
ETH, with a little help from @LidoFinance and @CurveFinance, can generate 12% yield. But where does this yield come from?
Let's break down the ponzu recipe. 🧫👇
The Ethereum blockchain gives rewards to computers that validate transactions. If you hold ETH, you can validate transactions. The easiest way to do this is to use a service like Lido. The yield is currently ~6%. lido.fi
Normally when you stake your ETH, your ETH is locked up. Lido gives you staked ETH (stETH) tokens in return. This makes you ETH liquid and allows you to do stuff with them.
The public cloud makes it easy for anyone to start a software company—but at a cost—your margins now belong to AWS. Thread:👇
There are three ways of paying for software infrastructure: 1. have your customer pay for it (cheap) 2. build your own data center (somewhat costly) 3. rent from a public cloud (very costly)
The cheapest infra is no infra. This is the classic enterprise software model: the customer buys your sw to run on their own hw. Selling pure sw yields the highest margins in industry: 90%+.
1/ How James Cameron’s Terminator 2 predicted modern AI chips and sparked the debate on AI safety. An appreciation thread.👇
2/ This is the chip that powers the T-800. Based on its appearance and commentary from chief architect Miles Dyson, the movie makes three predictions about future processors: 1) neural net acceleration 2) multi-core design 3) 3D fabrication.
Let’s look at these claims.
3/ Among the many technologies Cameron could have picked for Terminator, neural network processor was spot on. Neural net is the breakout technology of the past decade. As of 2020, there are ~100 companies building neural net processors with annual revenues exceeding $5 billion.