We've been asked this question for months. It highlights our contrarian business model: aligning our incentives with our users to build the best coding agent.
The question reveals the trap every AI coding tool falls into.
They all make money the same way: reselling AI inference at a markup. Buy wholesale, sell retail. The gas station that also built your car.
When you profit from AI usage, you're incentivized to either:
>charge more per token than it costs
>hide actual usage behind confusing credits
>route to cheaper models without telling users
Every decision optimizes for margin, not capability.
The math is brutal.
One power user coding with Claude can burn $500/day in AI costs. On a $200/month subscription.
The only options: limit usage or go bankrupt. That's why "unlimited" plans keep getting limited.
We built Cline differently.
You provide your own API keys, use any model, and pay the actual cost.
We earn from enterprise features requested by organizations -- such as team management, access controls, and audit trails -- not by marking up your AI usage.
This architecture makes betrayal impossible.
We can't throttle you in code you can read.
We can't route to cheaper models -- you choose.
We can't create artificial scarcity -- your limits are what you set.
When AI inference isn't our business model, our only path to success is making Cline more capable.
Not finding ways to give you less.
Not optimizing for margins.
Just building the best possible coding agent.
The future is clear: direct usage-based pricing.
When you can arbitrage a subscription in a commodity market, the market wins. Every AI coding tool will eventually converge on this model.
We just got there first.
2.7M developers have already figured this out.
F100 enterprises choose us because we're the only option that passes Zero Trust compliance. Your code never touches our servers.
Read why investors bet $32M on this approach and how we see the future unfolding:
The pattern is predictable: 1. launch with "unlimited" access 2. power users take you at your word 3. economics hit hard 4. add hidden limits 5. users discover limits when work stops
Subscription models for AI inference don't work. The math always wins.
Why? AI inference is a commodity like gas or electricity.
A $200/month power user can burn $500/day in AI costs. The provider loses $14,800/month per heavy user.
Only options: limit usage or go bankrupt. The market will arbitrage any free inference it is given.
They're abandoning RAG (via vector embeddings) for code exploration.
Why?
Code doesn't think in chunks, and it confuses the agent.
& codebase context that's gathered like a senior developer would leads to better outcomes. 🧵
When you embed code, you're literally tearing apart its logic. A function call in chunk 47, its definition in chunk 892, the context that explains why it exists scattered across a dozen fragments.
Even sophisticated chunking struggles with this. Natural language has obvious boundaries (paragraphs, sentences). Code? Not so much.
When your AI coding agent goes off track, your instinct is to course-correct. To explain what you really meant. To clarify, rephrase, add constraints.
But you're swimming upstream. Each correction adds more context pollution.
There's a better way -- using message checkpoints 🧵
Research from Microsoft and Salesforce just confirmed what we've been seeing: LLMs experience a 39% performance drop when you try to guide them through multiple conversation turns versus giving complete context upfront.
This fundamentally changes how we should think about AI interactions.
Think about it: when the model takes a wrong turn, it's not working with a blank slate anymore. It's trying to reconcile your corrections with its own generated assumptions.
It's like trying to merge back onto a highway from a field. The context is already polluted.
Our new guide to the MCP starter pack demonstrates how to integrate web search, live documentation, and browser automation directly into your Cline workflow. Equip Cline with the right context at the right time. 🧵
1. Supercharge your AI's knowledge with the Perplexity MCP. Give Cline access to the entire web for up-to-date research, so you never have to leave your editor to find answers.
2. Keep your code updated with the Context7 MCP. Access the latest documentation for over 4,000 libraries to ensure your AI generates accurate and timely code.
You can use your subscription in Cline instead of paying per-token API pricing.
Here's how you can set it up 🧵
Setup is simple:
1. Install Claude Code following Anthropic's guide 2. In Cline: Settings API Configuration > Select "Claude Code" 3. Set the path to your path to Claude Code CLI (can be just "claude")
If you can't find the path to your Claude Code CLI, trying running one of these commands:
macOS / Linux: which claude
Windows (Command Prompt): where claude
Windows (PowerShell): Get-Command claude