Ethan Mollick Profile picture
Jul 27, 2022 13 tweets 12 min read Read on X
Data visualization inspiration thanks to DALL-E: how Rothko, Basquiat, Picasso, and Monet would create an academic chart. ImageImageImageImage
A few more sources of data visualization inspiration: Bar charts as stained glass in an old cathedral. As a page from the Voynich Manuscript. As ancient stone monoliths on a grassy plain. Made of great columns of fire in the sky at the end of the world. cdn.discordapp.com/attachments/10… ImageImageImageImage
Bar charts made out of cake. In the style of Klimpt. As a Persian rug. Out of writhing tentacles. ImageImageImageImage
Bar charts in the style of the a 1950s comic book, by Leonardo da Vinci, made of Jello, on a knight's shield ImageImageImageImage
Scientific diagrams created by Vincent van Gogh, in the style of the Egyptian Book of the Dead, made of smoke and fire, haunted by ghosts. ImageImageImageImage
Bar charts in the style of Magritte, a burning post-apocalyptic city, a Byzantine mosaic, an 80s punk album cover. ImageImageImageImage
Bar charts made of books. Bones. Charcuterie. Tiny fuzzy monsters. ImageImageImageImage
Bar chart by Lisa Frank, in a book of dread prophecy, traced by the masts of tall ships in a Turner painting, outlined by tornadoes. ImageImageImageImage
Bar chart as cave painting, as Brutalist architecture, drawn by Studio Ghibli, in a frame of a Wes Anderson movie. ImageImageImageImage
Bar chart as drawn by Dali, in the style of DALL-E (I asked it to create a bar chart in the style of AI), composed of (creepy) dollies, made out of dal. ImageImageImageImage
Bar charts in the style of Keith Haring. As a traditional Chinese landscape. Carved into the rock of an alien planet. As a Hieronymus Bosch painting. ImageImageImageImage
Bar charts as a Cézanne still life. As a scene in a Michael Bay movie. As a couture dress. Out of art deco furniture. (All of these are done in Midjourney, which I used for the first time yesterday!) ImageImageImageImage
Since people keeping asking for these, here are all the images I posted (plus some leftovers), maximum resolution, under creative commons attribution license. Enjoy! drive.google.com/drive/folders/… Image

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Ethan Mollick

Ethan Mollick Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @emollick

Mar 24
As someone who has spent a lot of time thinking and building in AI education, and sees huge potential, I have been shown this headline a lot

I am sure Alpha School is doing interesting things, but there is no deployed AI tutor yet that drives up test scores like this implies. Image
I am not doubting their test results, but I would want to learn more about the role AI is playing, and what they mean by AI tutor, before attributing their success to AI as opposed to the other dials they are turning.

AI tutoring is still being understood
Google has been doing a lot of work on fine-tuning Gemini for learning, and you can see a good overview of the issues and approaches in their paper (which also tests some of our work on tutor prompts). arxiv.org/abs/2412.16429Image
Image
Read 4 tweets
Mar 15
I suspect that a lot of "AI training" in companies and schools has become obsolete in the last few months

As models get larger, the prompting tricks that used to be useful are no longer good; reasoners don't play well with Chain-of-Thought; hallucination rates have dropped, etc.
I think caution is warranted when teaching prompting approaches for individual use or if training is trying to define clear lines about tasks where AI is bad/good. Those areas are changing very rapidly.
None of this is the fault of trainers - I have taught my students how to do Chain-of-thought, etc. But we need to start to think about how to teach people to use AI in a world that is changing quite rapidly. Focusing on exploration and use, rather than a set of defined rules.
Read 4 tweets
Mar 8
“GPT-4.5, Give me a secret history ala Borges. Tie together the steel at Scapa Flow, the return of Napoleon from exile, betamax versus VHS, and the fact that Kafka wanted his manuscripts burned. There should be deep meanings and connections”

“Make it better” a few times… Image
It should have integrated the scuttling of the High Seas Fleet better but it knocked the Betamax thing out of the park
Dang, Claude. This is just half the thing.

Full story here: docs.google.com/document/d/1-h…Image
Image
Read 5 tweets
Mar 4
🚨Our Generative AI Lab at Wharton is releasing its first Prompt Engineering Report, empirically testing prompting approaches. This time we find:
1) Prompting “tricks” like saying “please” do not help consistently or predictably
2) How you measure against benchmarks matters a lot Image
Image
Using social science methodologies for measuring prompting results helped give us some useful insights, I think. Here’s the report, the first of hopefully many to come. papers.ssrn.com/sol3/papers.cf…
This is what complicates things. Making a polite request ("please") had huge positive effects in some cases and negative ones in others. Similarly being rude ("I order you") helped in some cases and not others.

There was no clear way to predict in advance which would work when. Image
Read 4 tweets
Feb 25
The lack of benchmarks for writing, telling stories, persuasion, creativity, emotional intelligence, perceived empathy, and doing office work are...

(1) holding back AI advances, (2) hiding big differences between models & (3) obscuring how good these models are for real work
If you want to influence the future, now is the time to release a really good benchmark.
We are getting AIs optimized for coding, doing graduate level math, multiple choice exams, and also counting the r's in strawberry.
Read 4 tweets
Feb 16
The significance of Grok 3, outside of X drama, is that it is the first full model release that we definitely know is at least an order of magnitude larger than GPT-4 class models in training compute, so it will help us understand whether 1st scaling law (pre-training) holds up.
It is possible that Gemini 2.0 Pro is a RonnaFLOP* model, but we are only seeing the Pro version, not the full ultra.

* AI trained on 10^27 FLOPs of compute, an order of magnitude more than then GPT-4 level (I have been calling them Gen3 models because it is easier)
And I should also note that everyone now hides their FLOPs used for training (except for Meta) so things are not completely clear.
Read 4 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(