If I write a couple of KBs of data to the Biitcoin or Ethreum bloockchains that data still gets copied to every single active node, right?

Any estimates as to how much total disk space those 2KB take up worldwide?
Asking because evidently the idea of "storing data on the bloockchain" is a frequent point of confusion, I wonder if explaining how many copies that entails would help clarify things at all
In January 2021 there were an estimated 83,000 active full nodes, so presumably any data you write to the bloockchain gets duplicated 83,000 times? coindesk.com/tech/2021/01/2…
Also, is the Ethreum bloockchain significantly larger than the Biitcoin one at this point? Presumably the need to execute smart contracts impacts the resources needed to run a node there?
Aha! Sounds like pools have a big impact here in saving their members from needing a full copy of the whole thing

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Simon Willison

Simon Willison Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @simonw

18 Nov
I've had a bit of a breakthrough with this over the past couple of years: maintaining detailed progress notes in a GitHub issues comment thread has dropped my "getting back on track" time down to a fraction of what it was
The reason it takes 25 minutes to spin back up again is that you're holding a ton of stuff exclusively on your own memory - so write it down!
Something I've realized is that 90% of software engineering is research, not typing code - figuring out what the code needs to do, which APIs to use, how best to test it etc

So all of that research goes in issue comments. Here's my best recent example: github.com/simonw/s3-cred…
Read 5 tweets
17 Nov
One of the biggest productivity tricks I'm using in the Datasette ecosystem is continuous deployment of live demos - every time I push to Datasette (+ a few other repos) it deploys a demo of latest main - it's fantastic for both catching bugs and linking to from issue comments
I've been working on the datasette-graphql plugin today and the live demo at datasette-graphql-demo.datasette.io/graphql helped me catch a bug where JS files were loading in the wrong order, breaking things - a problem that didn't occur on my laptop
Here's the issue thread for that github.com/simonw/dataset…
Read 4 tweets
13 Nov
Datasette is four years old today! It's come a long way since that first release back in 2017 simonwillison.net/2017/Nov/13/da…
Progress on the project is pretty thoroughly documented by the datasette tag on my blog - 249 items and counting simonwillison.net/tags/datasette/
Something that really excites me about Datasette is that I can genuinely see myself still being thrilled to work on it ten years from now.

Projects like that are pretty rare - I feel very lucky to have found one that combines so many of my interests all in one package
Read 5 tweets
12 Nov
I presented a 90 minute workshop on Git scraping and how to explore the resulting data using @datasetteproj at #CodaBr21 (by @EscolaDeDados) today

Here are the exercises from the workshop, plus additional screenshots and notes that I added afterwards

docs.google.com/document/d/1TC…
@datasetteproj @EscolaDeDados To save attendees from having to get a working Python environment setup on their laptop, I instead encouraged them to use a free @gitpod account (gitpod.io) - I demonstrated each exercise in GitPod too

Cloud-based development environments are SO GOOD for tutorials
(I had planned to use GitHub Codespaces for this, but then realized that those are not freely available to non-paid users outside of the beta program yet)
Read 6 tweets
21 Aug
Here's a fun challenge: given an array of datetimes, what's the best way to plot those on a frequency graph over time?

They might all be on the same day, or they might be spread out over several years - so the challenge is automatically picking the most interesting bucket size An example of the kid of chart I want to produce
Looks like d3.bin().thresholds() is the answer I'm looking for observablehq.com/@d3/histogram
Yup, this works perfectly: observablehq.com/@simonw/my-twe… Histogram of my tweets over time, generated with the followi
Read 6 tweets
9 Aug
Is there a reliable way to tell search engine crawlers that a site hasn't been updated in X days so they don't need to re-crawl it?

Do they tend to believe the <lastmod> element in sitemap.xml ? And can I set that to apply to the whole site, not just an individual page?
Asking because tailing logs shows a vast amount of crawler traffic to Datasette instances that haven't seen any data changes in over a year - I may have to robots.txt block crawlers from them to save in costs, but I'd rather tell them "no point in crawling, nothing has changed"
Datasette currently has a plugin for configuring robots.txt, but I'm beginning to think it should be part of core and crawlers should be blocked by default - having people explicitly opt-in to having their sites crawled and indexed feels a lot safer datasette.io/plugins/datase…
Read 4 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Too expensive? Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Thank you for your support!

Follow Us on Twitter!

:(