I repeat: Easily produced science text that's wrong does not advance science, improve science productivity, or make science more accessible. I like research on LLMs but the blind belief in their goodness does a disservice to them and science. Here is an example from #ChatGPT 1/5
SMPL is actually short for Skinned Multi-Person Linear model. #SMPL is a popular 3D model of the body that's based on linear blend skinning with pose-corrective blend shapes. It's learned from 3D scans of people, making it accurate and compatible with rendering engines. 2/5
Despite what #ChatGPT thinks, it wasn't developed at Berkeley or the MPI for Informatics. It was developed in the @PerceivingSys department of the @MPI_IS (the Max Planck Institute for Intelligent Systems). Run it again and you'll get different answers every time. 3/5
These sorts of errors *will* propagate. Users looking for a quick advantage will feel it's smart to use such tools. I think more effort should be spent on detecting incorrect or fake science - whether generated by a computer or a person. This would actually help science. 4/5
Side note: Using LLMs to generate science text reveals just how little the current models actually understand about the world. With non-science text, it's easier to believe that they know a lot. One day, computers will do good science. Just not today. 5/5 smpl.is.tue.mpg.de
• • •
Missing some Tweet in this thread? You can try to
force a refresh
In the LLM-science discussion, I see a common misconception that science is a thing you do and that writing about it is separate and can be automated. I’ve written over 300 scientific papers and can assure you that science writing can’t be separated from science doing. Why? 1/18
Anyone who has taught knows the following is true. You think you understand something until you go to teach it. Explaining something to others reveals gaps in your understanding that you didn’t know you had. Well, writing a scientific paper is a form of teaching. 2/18
Your paper is teaching your reader about your hypothesis, problem, method, the prior work in the field, your results, and what it all means for future work. When you write up your work and find it challenging, this is typically because you don’t yet fully understand it. 3/18
With LLMs for science out there (#Galactica) we need new ethics rules for scientific publication. Existing rules regarding plagiarism, fraud, and authorship need to be rethought for LLMs to safeguard public trust in science. Long thread about trust, peer review, & LLMs. (1/23)
I’ll state the obvious because it doesn’t seem to be obvious to everyone. Science depends on public trust. The public funds basic research and leaves scientists alone to decide what to study and how to study it. This is an amazing system that works. (2/23)
But it only works if scientists and the public each uphold their part of the deal. The pressure on scientists today to publish has never been greater. Publication and citation metrics are widely used for evaluation. (3/23)
I asked #Galactica about some things I know about and I'm troubled. In all cases, it was wrong or biased but sounded right and authoritative. I think it's dangerous. Here are a few of my experiments and my analysis of my concerns. (1/9)
I entered "Estimating realistic 3D human avatars in clothing from a single image or video". In this case, it made up a fictitious paper and associated GitHub repo. The author is a real person (@AlbertPumarola) but the reference is bogus. (2/9)
Then I tried "Accurate estimation of body shape under clothing from an image". It produces an abstract that is plausible but refers to
Alldieck et al. "Accurate Estimation of Body Shape Under Clothing from a Single Image"
Who should be the last/senior author on a paper? How do you decide? What does being last entail? I get these questions a lot and it’s confusing because the last author is often a senior person, running a group & raising money. Do those things determine last authorship? No. (1/7)
The last author is ultimately responsible for the paper throughout the process including conception of the idea, writing, rebuttal, camera ready, talk, video, code, website, tweet, dataset, etc. They don’t do everything but make sure it all gets done. Like a conductor. (2/7)
They are responsible for the paper’s intellectual integrity. If there is a mistake or worse, it’s the last author who will take the blame. It is not to be taken lightly.The buck stops with them. If something goes wrong, they have to fix it. (3/7)
Avatars are central to the success of the #metaverse and #metacommerce. We need different #avatars for different purposes: accurate #3D digital doubles for shopping, realistic looking for #telepresence, stylized for fun, all with faces & hands. @meshcapade makes this easy. (1/8)
For on-line shopping, clothing try-on, and fitness, an avatar should be realistic – your digital twin. You need a true digital double to see how clothing will look in motion. But, creating avatars that are accurate enough for shopping is hard. (2/8)
Since it’s hard to 3D scan everyone, digital doubles must be created from a few images or a video. Existing methods require users to wear tight clothes and have cumbersome capture protocols. @Meshcapade uses a single image of a person in any pose, making creation easy. (3/8)
The 5 stages of rebuttal grief. (1) Denial
The reviewers totally misunderstood my paper. The review process is broken. R1 was clearly a student who has never reviewed before. R2 doesn’t know what they are talking about. R3 hates me.
(2) Anger
I’m going to withdraw my paper. I’ll submit it somewhere else where other people will love it. I hate this conference and this field. The whole process is broken. Reviews are random.
(3) Bargaining
I’ll explain to the reviewers why they are so mistaken. I’ll convince them that my paper is great and that they are idiots. My reasoning will be so powerful, that they will be swayed and will accept my paper.