The more I've been using it (since Saturday, when I had an MVP), the more I realize that this kind of a tool is probably my best shot at building a Memex, a system that knows about and lets me search through my entire landscape of knowledge — theatlantic.com/magazine/archi…
I've probably performed ~100 searches for various names, ideas, memories, blogs, and other random things in the last week, and the most interesting thing is how searching for one thing helps me stumble into some unexpected insight or memory from my past. Creative randomness.
Lastly, great search and recall is a centerpiece of the "incremental note-taking" concept I discussed last week — monocle.surge.sh/?q=incremental…
Monocle is a system that doesn't need me to take notes; it gathers knowledge by looking through my existing digital footprint.
I've spent a bunch of time on this this weekend so probably going to take a small break, but hopefully in the coming weeks and months I'll add a few more data sources to my search index:
- Browser history, YouTube watch history
- Reading list from Pocket
- Email (maybe?)
Lastly, a question I'm definitely expecting is "can I run this on my own data?"
Uhh... .probably not right now? The system is pretty custom-built for my setup. But if I like it, I might make a version that's open for other people to try ✌️
Hypothesis: information work is overwhelmingly bottlenecked on availability of high-signal context more than by correct inference over the context. If right, implies higher ROI-per-flop of context building over pure logical inference.
h/t @anandnk24
Also, virtually all of the valuable context is in the tail of the information distribution. h/t @paraga
@anandnk24 wow this blew up. i dont have a soundcloud but go check out @PatronusAI to make sure your AI systems are behaving correctly and go make some ai apps on @ValDotTown
had a chance last night to meet with some of the best minds in AI to discuss the most pressing challenge facing society today:
✨how to afford attending the ERAS TOUR ✨
after much discussion, we've arrived at a breakthrough, what we've termed the "Taylor Swift Scaling Laws" 👇
the Taylor Swift Scaling Laws (TS2L) take inspiration from Scaling Laws for transformer-based LLMs, and apply the same log-log regression methodology to model and understand components of Taylor's ticket prices.
dare I say, we may have found something equally impactful
some highlights from the paper, which is a joint (overnight) work between me, @jtvhk, Ashish, and Niki (+ GPT4, thanks Notion for the OpenAI credit)
Full write up hopefully coming soon, but I'm using cosmo-xl for text generation with my own prompt, retrieving from an in memory vector DB with sentence-transformers embeddings, and using @sendbluedotco for iMessage.
The fact that LLMs generate text is not the point. LLMs are cheap, infinitely scalable black boxes to soft human-like reasoning. That's the headline! The text I/O mode is just the API to this reasoning genie. It's a side effect of the training paradigm.
A vanishingly small slice of knowledge work has the shape of text-in-text-out (copywriting/Jasper). The real alpha is not in generating text, but in using this new capability and wrapping it into jobs that have other shapes.
Exploring the "length" dimension in the latent space of a language model ✨
By scrubbing up/down across the text, I'm moving this sentence up and down a direction in the embedding space corresponding to text length — producing summaries w/ precise length control (1/n)
Length is one of many attributes that I can control by traversing the latent space of this model — others include style, emotional tone, context...
Here's "adding positivity" 🌈
It's a continuous space, so attributes can all be mixed/dialed more precisely than by rote prompting
More to follow soon on how it works, but in brief:
- Built on a custom LM arch based on T5 checkpoints
- An "attribute direction" is found from unpaired examples of texts w/ and w/o that trait
- Simple vector math in latent space + decoding from the latent gets you this effect.
Good tools admit virtuosity — they have low floors and high ceilings, and are open to beginners but support mastery, so that experts can deftly close the gap between their taste and their craft.
Prompt engineering does not admit virtuosity. We need something better.
Tools like Logic, Photoshop, or even the venerable paintbrush can be *mastered*, so that there is no ceiling imposed by the tool for how good you can get to going from image in your mind -> output. Masters of these tools can wield tools as extensions of themselves.
For this to work, the tool has to present a coherent set of abstractions, and predictable behavior about how composing them will change the user's output. Prompt engineering is not predictable, and there are no coherent abstractions. It's all just gut feelings and copy-paste.