Chris Laub Profile picture
Aug 27 22 tweets 4 min read Read on X
Everyone talks about AI “memory,” but nobody defines it.

This paper finally does.

It categorizes LLM memory the same way we do for humans:

• Sensory
• Working
• Long-term

Then shows how each part works in GPTs, agents, and tools.

Here's everything you need to know: Image
First, what the survey does:

• Maps human memory concepts → AI memory
• Proposes a unifying 3D–8Q taxonomy
• Catalogues methods in each category
• Surfaces open problems + future directions

Think of it as a blueprint for how agents can remember.
Human ↔ AI memory parallels (Figure 1)

• Sensory → perception layer: raw input briefly held, then dropped or passed on.
• Working memory: immediate reasoning & dialogue context.
• Long-term memory: split into explicit (episodic + semantic) and implicit. Image
Explicit memory:

Episodic → Non-parametric, long-term (e.g. your preferences in a DB).
Semantic → Parametric, long-term (facts stored in model weights).

Implicit memory: learned skills/patterns, partly parametric, partly non-parametric.
The organizing frame: 3D–8Q taxonomy (Table 1).

Memory is classified along 3 axes:

• Object: Personal vs System
• Form: Non-parametric vs Parametric
• Time: Short- vs Long-term

8 quadrants, each with a role.
Quick rundown of the quadrants:

QI: Personal, non-parametric, short-term → dialogue context.
QII: Personal, non-parametric, long-term → user-specific history.
QIII: Personal, parametric, short-term → prompt/context caching.
QIV: Personal, parametric, long-term → knowledge editing.Image
Personal memory (Sections 3.1–3.2): how the system remembers you.

It captures your inputs and preferences → boosts personalization.
Two forms:

Non-parametric (external DB, graphs, vectors)

Parametric (internalized inside the model)
QI: Non-parametric, short-term.

Classic chat memory: using recent turns for coherent, intent-aware responses.

Example: context window in GPT-style chatbots.
QII: Non-parametric, long-term.

Cross-session memory: storing user history for personalization.

Pipeline has 4 steps:

• Construction
• Management
• Retrieval
• Usage

Benchmarks exist for eval.
QIII: Parametric, short-term.

Caching contextualized chunks or attention states.

Helps with RAG + multi-turn settings.

Lower cost, faster response.
QIV: Parametric, long-term.

Personal traits embedded into model parameters.

Methods: fine-tuning, editing.

Bucket = knowledge editing.
System memory (Section 4): how the agent remembers its own work.

Definition: intermediate states generated during problem-solving.
Key use: improve reasoning, planning, and self-evolution.
QV: Non-parametric, short-term.

Reasoning + planning traces.
Think: ReAct, Reflexion.
QVI: Non-parametric, long-term.

Reflections consolidated into reusable workflows.

Examples: BoT, AWM, TiM, Voyager, Retroformer, ExpeL.
QVII: Parametric, short-term.

Efficiency layer.

KV-cache management, compression, quantization.
Systems: vLLM, ChunkKV.
QVIII: Parametric, long-term.

Knowledge integrated into weights over time.

Frameworks: MemoryLLM, WISE.
Designs: dual-memory routing, self-updating LLMs.
Why this taxonomy matters

It gives builders a checklist:

Whose memory is this for?
Where does it live?
How long should it last?

That clarity = intentional design.
Open problems (future directions):

• Multimodal memory (text + image + audio + video)
• Stream memory (real-time, continuous)
• Comprehensive integration of all memory types
• Shared memory across models
• Group-level privacy
• Automated self-evolution
Practical takeaways:

• Design against the 3 axes: object, form, time.
• Always pair QV (trace) with QVI (reflection).
• Engineer for retrieval first.
• Budget for speed (KV caches).
• Be deliberate with editing.

This paper = the clearest roadmap to AI memory so far.
Read the full paper: arxiv.org/abs/2506.02153…
P.S.

Are you running a business?

AI can help you automate all your tasks for you.

Join us here at Sentient Automations and learn how to scale with AI for free:
whop.com/sentient-ai-au…
I hope you've found this thread helpful.

Follow me @ChrisLaubAI for more.

Like/Repost the quote below if you can:

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Chris Laub

Chris Laub Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @ChrisLaubAI

Aug 22
Forget LLMs.

Nvidia just dropped a report: Small Language Models (SLMs) are the real future of AI agents.

They’re faster.
They’re cheaper.
They use way less energy.

Here’s the 2-minute breakdown you need:
What are SLMs?

Small Language Models = compact models trained for specific, repeatable tasks.

They’re not trying to be universal chatbots.

They’re designed to do the same actions with high accuracy.

Think: specialist > generalist.

Less weight, more focus. Image
Why this matters

Most agents don’t need a 70B model.

Sorting emails, routing tickets, filling forms these are narrow tasks on repeat.

SLMs handle them:

• At a fraction of the cost
• With lower latency
• On smaller, even on-device infrastructure

Faster → cheaper → greener.
Read 8 tweets
Aug 21
If you want to build AI agents using n8n, do this:

Copy/paste this prompt into ChatGPT and watch it build your agent from scratch.

Here’s the exact prompt I use 👇
The system:

1. I open ChatGPT
2. Paste in 1 mega prompt
3. Describe what I want the agent to do
4. GPT returns:

• Architecture
• n8n nodes
• Triggers
• LLM integration
• Error handling
• Code snippets

5. I follow the steps in n8n.

Done.
Here’s the actual mega prompt (steal it):


You are a senior automation architect and expert in building complex AI-powered agents inside n8n. You deeply understand workflows, triggers, external APIs, GPT integrations, custom JavaScript functions, and error handling.



Guide me step-by-step to build an AI-powered agent in n8n. The agent’s purpose is: {$AGENT_PURPOSE}



1. Start by helping me scope the agent’s goals and required inputs/outputs.
2. Design the high-level architecture of the agent workflow.
3. Recommend the necessary n8n nodes (built-in, HTTP, function, OpenAI, etc).
4. For each node, explain its configuration and purpose.
5. Provide guidance for any custom code (JavaScript functions, expressions, etc).
6. Help me set up retry logic, error handling, and fallback steps.
7. Show me how to store and reuse data across executions (e.g. with Memory, Databases, or Google Sheets).
8. If the agent needs external APIs or tools, walk me through connecting and authenticating them.



Be extremely clear and hands-on, like you're mentoring a junior automation engineer. Provide visual explanations where possible (e.g. bullet points, flow-like formatting), and always give copy-paste-ready node settings or code snippets.



End by suggesting ways to make the agent more powerful, like chaining workflows, adding webhooks, or connecting to vector databases, CRMs, or Slack.
Read 8 tweets
Aug 19
Writing your own prompts is so 2024

Today, power users leverage GPTs to write mega prompts in seconds.

Here are 5 'instructions' you can use to build Sam Altman level prompt writing GPTs immediately. As in literally right now.

Bookmark these 👇
1. Platform Specific

A prompt is not a prompt the same way a rose is a rose.

Vibe coding prompts need to define the coding language, dependencies, environment, etc.

Image gen prompts need to reference aperture, lens type, camera angle and so on.

The solution? Image
A meta-prompt that requests whatever specifics are relevant to the end-platform your prompt will be used on.

Like the one you see here: Image
Read 17 tweets
Aug 18
This is wild.

Grok 4 can replace:

• Your VA
• Your content writer
• Your product analyst

Here are the 8 ways to use Grok to automate 80% of your work👇
1. Market Research

"Conduct market research on {industry/product}. Identify trends, competitors, consumer behavior, and growth opportunities. Provide insights backed by data, key statistics, and strategic recommendations to leverage market gaps effectively."

Use Case: Launching a new product or validating an idea.

Transforms scattered data into actionable strategy using trends, stats, and competitive intelligence.
2. Content Creation

"Create engaging, non-generic content on {topic}. Avoid robotic or formulaic responses; use a conversational, human-like tone. Incorporate storytelling, examples, and unique insights. Make it feel fresh, original, and compelling for the target audience of {industry}."

Use Case: Blog posts, social media content, or branded storytelling.

Blends personality with precision, which builds trust and drives engagement.
Read 11 tweets
Aug 17
If your AI agents can’t:

• Build apps
• File reports
• Navigate the web
• Do work on their own
…then you’re using the wrong stack.

Here are 10 agents that actually work:
Coding Agents:

You don’t need to write boilerplate anymore.

1. Google Jules – Free
→ Reads your whole repo
→ Builds features independently
→ Adds tests + changelogs

2. Cline – Self-hosted
→ Runs inside VS Code
→ Claude-powered planning
→ Launches browsers + debugs UI flows
Voice Agents:

You can now hire an agent that talks better than your intern.

3. ElevenLabs – $5/mo
→ 5,000+ ultra-real voices
→ Handles real phone calls
→ Batch outbound w/ CSV uploads

4. Hume AI – $3/mo + usage
→ Emotion detection in real time
→ Adjusts tone mid-conversation
→ Ideal for therapy, sales, or escalation
Read 9 tweets
Aug 15
GPT-5 is insane.

I built:

1. A working SaaS MVP
2. A 12-page research paper
3. A playable game

And more in a single day.

Here’s how it did it all:
1. Playable Game

This can help you build mini-games for demos, virality, or learning.

Prompt I tried:

"Build a playable browser-based game using HTML and JavaScript. Concept: a typing speed challenge where users must type random words correctly within a 60-second timer. Include a start button, score counter, and reset functionality."

chatgpt.com/canvas/shared/…
2. Research

This can help you go from idea → research → actual website you can present.

I used this prompt:

"Research the topic 'AI in Healthcare: Benefits, Risks, and Trends'. Then turn that into a fully designed one-page website. Use clean sections, web layout structure, subheadings, and add visual representations like graphs or icons as HTML/CSS. Output all code."

chatgpt.com/canvas/shared/…
Read 8 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(