1/ Just like "transactions per second" does not adequately reflect the real adoption of a blockchain network (as it doesn't take into account the economic value of each transaction)
The number of "requests fulfilled" for an oracle network does not accurately track adoption
2/ This is because most oracle networks operate using a decentralized data model where an on-chain reference contract is created to represent some value (e.g. the ETH/USD exchange rate)
Each update of these data feeds is considered a "request fulfilled"
However, the update frequency of each data feed is determined by the transaction throughput of the underlying blockchain where data is being delivered
It is not the oracle network which is the bottleneck here, it's the on-chain environment
3/ For example, widely used blockchains like Ethereum (limited to 15 TPS) have limited bandwidth which caps the maximum update frequency of any data feeds operating on that network
This is true for any oracle network on such blockchain networks
4/ Layer 2 solutions and alternative blockchains that support 10,000+ TPS have much lower transaction fees, supporting data feeds which update at a much higher frequency
However, the majority of smart contracts which require external data are on widely used chains like Ethereum
5/ For this reason, Chainlink is a blockchain agnostic framework for building oracle networks that natively deliver data to Ethereum, Polygon, BSC xDai, HECO, Starkware, and more
Chainlink oracle networks operate at the native speed of the underlying blockchain receiving data
6/ For example, Chainlink price feeds operating live on @StarkWareLtd have an update latency of less than second and consume next to zero transaction fees thanks to the use of an L2 zkRollup
6/ Therefore when an oracle project claims their solution is "super fast, low cost, and fulfills a bazillion requests", what they actually mean is that they are putting data onto high throughput chains and ignoring widely used chains to support their misleading speed narrative
7/ There are even oracle projects which have created their own blockchain with extremely high, long-term hardware requirements, making it even easier for them to inflate this "requests fulfilled" metric to mimic adoption (regardless if that data is actually being used)
8/ This data can't even be used until it is actively bridged to the blockchain networks where externally-connected smart contracts operate
Usually, only a small fraction of these "requests fulfilled" are actually bridged to other chains
7/ Such projects further the deception by purposely comparing the frequency of updates of Chainlink data feeds *on Ethereum* to their oracle networks delivering data to a blockchain with no smart contracts
A total apple to oranges comparison
8/ The number of "requests fulfilled" of an oracle network tells nothing about the amount of value being secured, the total number of integrations, the quality of the data, the market coverage provided, the economies of scale pricing model, and certainly not the level of adoption
9/ Yes this is a subtweet of a project who does exactly this by disseminating this false narrative through their disinformation marketing team
However I also seeing people promote this metric in a genuine way but it really doesn't tell the whole story
10/ With the release of Chainlink OCR, node repsonses are batched into a single transaction, which is a good thing for scalability, but actually lowers the number of "requests fulfilled", making this metric even more useless for comparing different oracle solutions
11/ "Requests fulfilled" isn't the only metric that competiting "LINK killers" try to manipulate to make their project more appealing, another is "Total Value Locked" and decentralization of the network as a whole, but I'll save that for another thread
• • •
Missing some Tweet in this thread? You can try to
force a refresh
Expanding on this, colors are basically just a mental construct, it is how your brain interprets specific wavelengths, my colors may not be the same as yours
If one of your cones doesn't work, your perception of "reality" is skewed, though not really any more or less"true"
Pain is just your brain's response to damaged nerve cells which send electrical signals that are routed by your peripheral system to the central nervous system (specifically your nrian) to stop whatever action causes damage
Your brain simulates "pain" to incentivize the rest of "you" to fix this problem, if this didn't happen (no pain threshold), humans would die far more easily without realizing it en.m.wikipedia.org/wiki/Congenita…
3/ What makes the combination of @Chainlink + @Arbitrum truly unique is regarding that first image
Chainlink oracle nodes can run Arbitrum L2 Validators and execute smart contracts off-chain in a highly scalable manner (270x more throughput in some cases) medium.com/offchainlabs/s…
The "team is dumping" narrative is a meme that needs to be put to rest
These tokens are being used to bootstrap and subsidize the growth of the most widely used oracle network
Reinvestment into a network's growth is a good thing, because projects that don't, stagnate and die
The token movements you see are used for a number of initiatives including:
- Subsidizing oracle networks to bootstrap a secure network of nodes
- Paying out community grants and blockchain integration grants
- Acquiring new technologies like DECO and funding security audits
- Hiring more devs to create software like OCR that lower gas costs by 90%
- Hiring more engineers to increase the rate integrations can occur
- Hiring more researchers to formalize the economics of staking
- Hiring geniuses like Ari Juels (RSA, IC3) and Ben Chan (BitGo, WBTC)
The new $LINK vault on @iearnfinance is a very welcome addition
Currently generates 70% APY
Zero impermanent loss risk
Lowers slippage of large LINK trades
Uses Chainlink oracles (Synthetix)
DeFi composability sandwich
2/ Off-Chain Reporting (OCR) is where the collection of oracle responses is done off-chain
A single transaction is submitted on-chain containing every oracle's individual observation and signature
Median value is determined on-chain because the observations are already sorted
3/ The reason data isn't medianized off-chain before delivery in the OCR model is because the individual ECDSA signatures can't be used without the original pre-medianized data to compare against
We need the cryptographic security of signatures or observations could be falsified