Asuka🎀Redpanda Profile picture
May 14 8 tweets 2 min read Read on X
visited my uncle in shenzhen. he’s a gpu smuggler.
he handed me this modified 5090 turbo and said:
"future of AI inference. 32GB each, 8 cards, 256GB total VRAM, under $30k. huaqiangbei doesn’t wait for nvidia."
huaqiangbei is really wild.💀Image
Image
here’s what he told me: HGX servers are designed for training huge AI models—power-hungry, liquid-cooled, and crazy expensive. But for inference (running those models), it’s a different game: → You don’t need as much compute → You just need enough VRAM to fit the model
That’s why many AI infra builders use traditional x86 + PCIe servers:
• cheaper
• flexible
• easy to scale horizontally
But there’s a problem: consumer GPUs like 4090/5090 are big and awkward—2.5 to 4 slots wide.
Enter the blower-style card: double-slot, front-to-back airflow, server-friendly.
Each generation has one. But NVIDIA hates them.
Why? Because a rack full of 4090 blowers replaces an H100 server at 1/10 the cost.
NVIDIA cripples gaming cards on purpose:
🚫 No NVLink after 3090
🚫 Max 24GB VRAM
🚫 No official blower 4090/5090
So if you want dense GPU inference, you either go broke... or go underground.
In Huaqiangbei, engineers reverse-engineered the blower design.
Now they mass-produce 4090 blowers, unofficial and off-NVIDIA’s radar.
They're shipping globally, and account for 90%+ of all 4090 blowers in the wild.
This has accidentally made the 4090 the go-to choice for inference servers because it’s crazy cost-effective. Sure, it doesn’t have NVLink, but with some software wizardry, you can still pool the VRAM—24GB times 8 cards gives you 192GB total—to run big models under 200 billion parameters, or even FP4-quantized high-parameter models.
Huaqiangbei takes it even further. They’ve figured out a way to mod 4090s to have 48GB of VRAM. That means you can build an inference server with 384GB of VRAM for under 50k. And right now all those huaqiangbei GPU bros are busy producing blower-style 5090 cards, which my uncle believes will become the next big thing for affordable, high-performance inference servers.Image

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Asuka🎀Redpanda

Asuka🎀Redpanda Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @VoidAsuka

Aug 2, 2023
今天起正式打工三周年啦。😭
刚去北京的时候一个人也不认识,周末只能漫无目的地散步,碰巧走进的万圣书园,为了避免形成自己的信息茧房决定拥抱随机性,去书店闭着眼睛拿了两本书,一本书是关于山本耀司,一本书是康定斯基。 1/n
两个人都是学法律出生的,山本耀司是庆应大学的法律系,毕业后还是想做衣服,于是又读了一个服装设计的专门学校,而康定斯基在莫斯科大学做法律学讲师一直到三十岁,三十岁去了慕尼黑学习画画,现在都是历史上闪闪发光的艺术家。高中读到毛姆《月亮与六便士》时候便感觉这种事也会发生在我身上。2/n
所以当我在书店看完那两本书的时候,便有一种非常强烈的宿命感。那个时候我也从事着和我本科专业物理学毫无关联的程序员岗位,想着自己攒够三年的钱,就一定要去武藏野或者多摩或者日艺学习艺术。🥲如今已经满三年了,而我好像离这个宿命越来越远了。3/n
Read 4 tweets
Dec 14, 2022
怎么短时间内快速英语技术面试水平?面对全英的技术面试,大部分人想到的方法是系统学习提高自己的口语,比如上cambly等平台,或者请便宜的一对一菲律宾外教等。而我认为这种方法效率非常低。1/n
因为传统的练习口语平台涉及到的话题太多太泛,包括日常、旅行、哲学、天气、地理……而这些内容是你会和技术面试官谈的嘛?你会和他分享你昨天去的餐厅哪道菜最好吃嘛??等入职后,和同事做一些small talk是有必要的,但得先有offer再说。2/n
要想快速搞定一场英文技术面试,最好的办法就是多mock,mock次数越多你就越熟练。一场英文技术面试涉及到的话题翻来覆去无非就是:1.个人背景; 2.项目经验deep dive; 3.刷题(边说要边用英文解释思路);4.系统设计;5.行为面试;6.岗位相关基础知识。3/n
Read 7 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(