The new iOS app from @runwayml featuring #Gen1 is 🔥 and now widely available!
- Turn anything into everything using your phone's camera
- Revive forgotten videos on your camera roll
- Effortlessly transfer assets to your Runway account without airdropping!
He calls it KERNEL, and it's transformed how his entire team uses AI.
Here's the framework:
----
K - Keep it simple
Bad: 500 words of context
Good: One clear goal
Example: Instead of "I need help writing something about Redis," use "Write a technical tutorial on Redis caching"
Result: 70% less token usage, 3x faster responses
----
E - Easy to verify
Your prompt needs clear success criteria
Replace "make it engaging" with "include 3 code examples"
If you can't verify success, AI can't deliver it
My testing: 85% success rate with clear criteria vs 41% without
----
R - Reproducible results
Avoid temporal references ("current trends", "latest best practices")
Use specific versions and exact requirements
Same prompt should work next week, next month
94% consistency across 30 days in my tests
----
N - Narrow scope
One prompt = one goal
Don't combine code + docs + tests in one request
Split complex tasks
Single-goal prompts: 89% satisfaction vs 41% for multi-goal
----
E - Explicit constraints
Tell AI what NOT to do
"Python code" → "Python code. No external libraries. No functions over 20 lines."
Constraints reduce unwanted outputs by 91%
----
L - Logical structure Format every prompt like:
Context (input)
Task (function)
Constraints (parameters)
Format (output)
----
Real example from my work last week:
Before KERNEL: "Help me write a script to process some data files and make them more efficient"
Result: 200 lines of generic, unusable code
After KERNEL:
Task: Python script to merge CSVs
Input: Multiple CSVs, same columns
Constraints: Pandas only, <50 lines
Output: Single merged.csv
Verify: Run on test_data/
Result: 37 lines, worked on first try
----
Actual metrics from applying KERNEL to 1000 prompts:
First-try success: 72% → 94%
Time to useful result: -67%
Token usage: -58%
Accuracy improvement: +340%
Revisions needed: 3.2 → 0.4
----
Advanced tip from this user:
Chain multiple KERNEL prompts instead of writing complex ones.
Each prompt does one thing well, feeds into the next.
They've open-sourced bitnet.cpp, a blazing-fast 1-bit LLM inference framework optimized for CPUs.
This is a major step forward for running large models locally, without expensive GPUs or cloud costs.
Demo app + repo + paper in 🧵 ↓
1/
Key highlights:
→ Achieve up to 6x faster inference with 82% lower energy consumption
→ Run 100B parameter models directly on x86 CPUs
→ Leverage ternary weights (-1, 0, +1) and 8-bit activations to dramatically reduce memory usage
2/
Alongside this, Microsoft also released BitNet b1.58 2B4T, the first functional open-source model using just 1.58 bits for weights while maintaining strong benchmark performance.
If you care about efficient AI at scale, this is worth a look!
Most people don’t realize ChatGPT has hidden operators that can totally change its answers.
Here's the ultimate 32-shortcut cheatsheet for sharper prompting! 🤯
Add one to the start.
Example: /ELI5: [topic] → explain this topic like I’m five
Full list + sheet in 🧵 ↓
👇 Here’s a complete list of ChatGPT operators:
/ELI5 is used to explain as if to a 5-year-old.
/TLDL summarizes a very long text in a few lines.
/STEP-BY-STEP lays out reasoning step by step.
/CHECKLIST turns a response into a checklist.
/EXEC SUMMARY gives a quick executive-style summary.
/ACT AS makes ChatGPT speak in a specific role.
/BRIEFLY forces a very short answer.
/JARGON asks to use technical vocabulary.
/AUDIENCE adapts the response to a chosen audience.
/TONE changes the tone (formal, funny, dramatic, etc.).
/DEV MODE simulates a raw, technical developer style.
/PM MODE gives a project-management perspective.
/SWOT produces a strengths/weaknesses/opportunities/threats analysis.
/FORMAT AS enforces a specific format (table, JSON, etc.).
/COMPARE puts two or more things side by side.
/MULTI-PERSPECTIVE shows several points of view.
/CONTEXT STACK keeps multiple layers of context in memory.
/BEGIN WITH / END WITH forces starting or ending with something.
/ROLE: TASK: FORMAT: explicitly defines the role, the task, and the expected format.
/SCHEMA generates a structured outline or a data model.
/REWRITE AS: rephrases in a requested style.
/REFLECTIVE MODE prompts the AI to reflect on its own answer.
/SYSTEMATIC BIAS CHECK asks to identify biases.
/DELIBERATE THINKING forces slower, more thoughtful reasoning.
/NO AUTOPILOT forbids superficial, autopilot responses.
/EVAL-SELF asks for a critical self-evaluation of the response.
/PARALLEL LENSES examines from several angles in parallel.
/FIRST PRINCIPLES rebuilds from fundamental basics.
/CHAIN OF THOUGHT shows intermediate reasoning.
/PITFALLS identifies possible traps and errors.
/METRICS MODE expresses answers with measures and indicators.
/GUARDRAIL sets strict boundaries not to cross.
Spotted this gem from Kieran Flanagan (@searchbrat) on LinkedIn.
It’s an o3 prompt that scores your page using Ogilvy’s copywriting playbook and gives clear steps to make it better! 🔥
Here’s how it works 👇
------
This prompt will:
- Take a URL
- Extract the web copy
- Run it through 15 principles based on David Olgiviy's work
- Score the web copy across those principles
- And suggest edits on how to reach a score of 100.
------
📝 Here's the prompt:
"You are an advertising strategist trained in David Ogilvy’s principles.
Task:
1. Visit the user-provided URL. 2. Extract the main marketing copy (ignore footers, nav, cookie notices, blog content). 3. Score the copy out of 100 using the 15 Ogilvy-inspired principles (each ~6.7 points). 4. Provide a detailed score breakdown. 5. Identify the top 3 improvement areas. 6. Suggest edits to improve the score. 7. Rewrite the copy to achieve 100/100.
---
### 15 Scoring Criteria:
1. **Product Positioning** — Is the offer clear? What is it, who is it for, and why it matters? 2. **Unique Benefit** — Is there a strong, specific benefit? 3. **Headline** — Is it clear, specific, curiosity-driving, or benefit-led? 4. **Reader-Focused** — Is the copy centered on the reader's needs, not the brand? 5. **Clear Tone** — Is it plainspoken, not vague or gimmicky? 6. **Simple Language** — No jargon, easy to understand? 7. **Evidence** — Are there facts, stats, testimonials, or proof? 8. **Emotion/Story** — Is there emotional or narrative appeal? 9. **Structure** — Is it skimmable and well-formatted? 10. **Call-to-Action** — Is the next step obvious and compelling? 11. **Visuals/Captions** — If present, do they reinforce the message? 12. **Testability** — Can parts be A/B tested or measured? 13. **Length** — Is it appropriate for product complexity? 14. **Attention-Grabbing** — Does it hook early? 15. **Repetition** — Are key ideas or benefits repeated effectively?