researcher @google; serial complexity unpacker;
https://t.co/Vl1seeNgYK
ex @ msft & aerospace
10 subscribers
Jul 30 • 4 tweets • 2 min read
Programming Languages used to be designed as powerful as possible.
Maximum possible utility given hardware constraints.
The pro move is to choose the *least* powerful, non-Turing-complete solution.
The entire web exists because of the Principle of Least Power:
Don’t take my word for it. Tim Berners-Lee (inventor of HTML, HTTP, etc) had this to say:
“the less powerful the language, the more you can do with the data...”
HTML is purposefully *not* a real programming language.
The constraint pushed innovation to data processing.
Jul 29 • 4 tweets • 2 min read
When you make a Bank ACH transaction, it’s literally just an SFTP upload.
Sent as a NACHA file, it's 940 bytes of ASCII text.
Bank-to-Bank transactions cost ~0.2 cents. As long as it travels via encrypted tunnel; it’s compliant!
Here’s how the quirky system works:
Chase offers a sample NACHA file to look at.
Notice the rows padded with 9s. It’s an artifact of a 1970s rule about magnetic tape, "always fill the block".
To this day, total line count *must* be a multiple of ten; otherwise the bank will drop the transaction.
Jul 28 • 4 tweets • 2 min read
Intel’s not doing so hot lately. Meanwhile vendors are killing it at the RISC-V Summit in China.
One CPU got a specint2006/GHz rating of 10.4/GHz!
To put it in perspective, a i7-4790k (Haswell) scores 8.1/GHz.
RISC-V is hitting high-end desktop territory FAST:
No one realizes how quickly a CPU supply chain; completely independent of western IP is progressing:
Not long ago RISC-V performance was a joke.
Now it’s trading blows with x86 and high end ARM!
In another generation or two it’s going to be a *serious* contender.
Jul 22 • 5 tweets • 3 min read
Fading out audio is one of the most CPU-intensive tasks you can possibly do!
Values that get close (but not quite) zero, hit an underflow gap known as "Subnormal" range.
It’s a mathematical conundrum so tricky, both x86 and ARM made special CPU instructions just to handle it!
In computer science, mathematical real numbers are approximated by floating-point representations.
For a single 32-bit float, the smallest “normal” positive we can hold is 1.17 × 10^-38
Tiny differences could get rounded off, leading to later Divide by Zero errors!
Jul 21 • 4 tweets • 2 min read
Windows is one massive (private) Git repo.
When I was at MS, the Windows Source had around ~3k PRs a day!
Regular Git didn’t scale to those levels at the time.
Internally there was a progression from Git -> GVFS -> Scalar -> merge back to Git. Here's how it worked:
Cloning Defender’s antivirus repositories alone would take a full day!
You may be wondering how we dealt with merge conflicts.
Teams were heavily siloed, with most of the effort put up front on the build process.
Syntax and such was checked with local builds before pushing.
Jul 14 • 5 tweets • 3 min read
Microsoft took ~30 years to be compliant with the C++ standard!
Seriously. From 1993 to 2020, MSVC’s preprocessor wasn’t feature-complete.
Code that compiles perfectly on Linux often broke.
Hold your judgement; there's some interesting historical nuance:
A tech race in the 80s led to unfixable debt.
Official standards wouldn’t exist until 1998.
MS engineers made “best guesses”, but they were competing with others (Borland, Watcom) for the C++ compiler market.
"We'll clean it up after capturing market share".
Jul 10 • 4 tweets • 2 min read
There’s a cursed C++ competition where programmers try to create the largest possible error message.
Finalists created ~1.5GB of error messages from just 256 bytes of source.
Preprocessor exploits were so easy, they had to create a separate division! Here's my favorites:
One contestant experimented with C-Reduce; a way to search for C++ programs that create unusual compiler behavior.
Maximizing the fitness function for error as a reward, no templates, no pre-processor hacks.
Just nested parenthesis causing exponential error output!
Jul 9 • 5 tweets • 3 min read
What happens when you freeze a process *perfectly*? RAM, VRAM, network, everything.
Imagine:
- Live-migrations of LLM training jobs
- time-travel debugging
- Surgical repairs of a crash moments before segfault
It’s called CRIU, and it’s already here:
It starts with a mad Russian.
At least, that’s what a lead linux kernel dev called it:
“a project by various mad Russians to perform c/r mainly from userspace, with various oddball helper code...I'm less confident than the developers that it will all eventually work”
Jul 7 • 4 tweets • 3 min read
Humans live at 10 bits per second.
The brain takes in ~11 million bits per second of sensory data, yet the inner conscious workspace is massively compressed.
Most people speak at ~40 b/s. How can we speak faster than we can think?
It's all about error correction:
Speech may exceed the cognitive speed limit, but most of the bits are redundant.
Language is designed to withstand noise and mis-hearing. The 40bit “raw rate” is predictable from context.
A Caltech study shows the effective payload collapses to <13 b/s when stripped.
Jul 5 • 4 tweets • 2 min read
Whole-home lithium power used to be a rich man’s game.
Now it’s “high-end graphics card” territory.
This is a $2500, lithium polymer battery that would power an entire US residential house for >24hr.
China is *crushing* it on kilowatt hours per dollar.
Let’s put it into perspective. That battery is 2x the kWh of a tesla powerwall 3.
Each powerwall will set you back $15k a piece.
Residential battery setups usually cost $1000 per kWh.
This is $80 per kWh.
Jul 4 • 4 tweets • 2 min read
Ubuntu’s next version won’t work on 90% of current RISC-V computers.
Got a raspberry-pi style RISC-V board? Yeah, not gonna work.
This isn't a bad thing.
It’s actually a genius move that will force the industry in the right direction; helping consumers in the process.
Here’s the problem with a debian-style “we support everything” mentality:
Many board makers slap cheap, borderline out-of-spec Chinese RISC-V CPUs onto a PCB and call it a day.
The bare minimum for a functional OS.
Many are riddled with hardware-level vulnerabilities (c908)!
Jun 20 • 5 tweets • 3 min read
A lone Boston coder rewrote BIOS in 1984.
IBM wanted to sue. The programmer's clever loophole became the model for legally defensible reverse engineering.
You’ve probably been booting his descendants ever since. This is how Phoenix Technologies got away with it:
The case Apple v. Franklin held that BIOS could be protected by copyright.
IBM already crushed other competitors citing the case law.
Phoenix created a bizarre ruse. What if an engineer had never *seen* IBM’s specification…and thus came up with the idea organically?
Jun 12 • 4 tweets • 3 min read
Grep is actually an acronym; originally started as a ed command:
:g/<re>/p
G - global search. <re> - Regular Expression, P - print.
It was invented to run Natural Language Processing on The Federalist Papers; the precursor to the US Constitution. Yes, really.
Lee McMahon, computer scientist at Bell Labs, was working to clarify the authorship of The Federalist Papers.
85 essays, all published under a pseudonym, it’s been a historical puzzle for a century+
The trick? De-anonymization by isolating word-frequency statistics.
Jun 3 • 4 tweets • 2 min read
A squadron of F22’s was once taken out by an imaginary line.
On a mission to Japan, an unforeseen software bug occurred crossing the international date line. Longitude swaps from W179.99 to E180 degrees.
Navigation, comms, and even fuel management went down!
This wasn’t a simple "turn it on and off again” fix; something was seriously wrong. Reboots weren't helping.
According to Maj. Gen. Sheppard:
“…all systems dumped and when I say all systems, I mean all systems...they could have been in real trouble."
May 27 • 4 tweets • 2 min read
Want to recognize a song from just a few seconds of distorted audio?
Use Constellation Maps.
The math is brilliantly simple.
With just a handful of bytes; discarding 99% of the waveform, you can recognize a unique fingerprint across hundreds of millions of tracks.
First, chop up the audio into few-second windows.
Take an FFT of the waveform, then extract the local peaks. Each maximum becomes a “star” on an xy plot of time vs frequency.
Pair nearby stars into clusters and hash the result. Boom, a noise-resistant fingerprint.
May 23 • 4 tweets • 3 min read
I miss the insanity of 80s processor design.
Intel’s iAPX 432 was a “micromainframe”.
It had no general purpose registers, supported object orientation *directly*, and performed garbage collection on-chip.
It was also 23x slower than an 8086. Here's why it failed.
Intel targeted Ada so aggressively that C support was an afterthought.
Problem was, particularly at the time, the Ada compiler was extremely untuned and immature.
Scalar instructions were basically never used; *everything* was huge object-oriented calls.
May 22 • 5 tweets • 3 min read
NTIRE is the coolest conference you’ve never heard of.
Deleting motion blur? Sure.
Night Vision? No problem.
Every year, labs compete on categories like hyperspectral restoration, satellite image enhancement, even raindrop removal (think car sensors)! Some highlights ->
Low-light enhancement is always popular.
Retinexformer, shown here got 2nd place in the 2024 contest.
A *TINY* transformer-based model, it runs in about 0.5 seconds for a 6K image on a single 3090. Only 1.6M parameters (<2MB weights at INT8)!
May 21 • 5 tweets • 3 min read
What if an OS fit entirely inside the CPU’s Cache?
Turns out we’ve been doing it for decades.
CNK, the OS for IBM’s Blue Gene Supercomputer, is just 5,000 lines of tight C++.
Designed to “eliminate OS noise”, it lives in the cache after just a few milliseconds of boot.
Kernels that “live” in the cache are common for HPC.
Cray’s Catamount microkernel (~2005) used a similar method for jitter free timing.
Huge Pages, Statically Mapped Memory, and a lack of scheduling are all typical aspects of these systems.
What about the modern era?
May 13 • 5 tweets • 3 min read
TDP (Thermal Design Power) of CPUs is a garbage metric that misleads consumers.
In the Pentium era, a 89W TDP meant just that; expect to dissipate 89W of heat in the worst case.
With Alder Lake, a 125W CPU can draw ~241W indefinitely!
Here's the goofy math:
CPU’s didn’t really know how to idle until the early 2000s. They just kinda ran full bore all the time.
With the introduction of C-States, various parts of the processor could be shut down, saving power when the computer was doing nothing.
Of course, this was HUGE for laptops.
May 12 • 6 tweets • 4 min read
What if humanity forgot how to make CPUs?
Imagine Zero Tape-out Day (Z-Day), the moment where no further silicon designs ever get manufactured. Advanced core designs fare out very badly.
Assuming we keep our existing supply, here’s how it would play out:
Z-Day + 1 Year: