We're seeing multiple folks in #NLProc who *should know better* bragging about using #ChatGPT to help them write papers. So, I guess we need a thread of why this a bad idea:
>>
1- The writing is part of the doing of science. Yes, even the related work section. I tell my students: Your job there is show how your work is building on what has gone before. This requires understanding what has gone before and reasoning about the difference.
>>
The result is a short summary for others to read that you the author vouch for as accurate. In general, the practice of writing these sections in #NLProc (and I'm guessing CS generally) is pretty terrible. But off-loading this to text synthesizers is to make it worse.
>>
2- ChatGPT etc are designed to create confident sounding text. If you think you'll throw in some ideas and then evaluate what comes out, are you really in a position to do that evaluation? If it sounds good, are you just gonna go with it? Minutes before the submission deadline?>>
3- It breaks the web of citations: If ChatGPT comes up with something that you wouldn't have thought of but you recognize as a good idea ... and it came from someone else's writing in ChatGPT's training data, how are you going to trace that & give proper credit?
>>
4- Just stop it with calling LMs "co-authors" etc. Just as with testifying before congress, scientific authorship is something that can only be done by someone who can stand by their words (see: Vancouver convention).
5- I'm curious what the energy costs are for this. Altman says the compute behind ChatGPT queries is "eye-watering". If you're using this as a glorified thesaurus, maybe just use an actual thesaurus?
6- As a bare minimum baseline, why would you use a tool that has not been reliably evaluated for the purpose you intend to use it for (or for any related purpose, for that matter)?
/fin
p.s.: How did I forget to mention
7- As a second bare minimum baseline, why would you use a trained model with no transparency into its training data?
@willknight The 1st is somewhat subtle. Saying this ability has been "unlocked" paints a picture where there is a pathway to some "AI" and what technologists are doing is figuring out how to follow that path (with LMs, no less!). SciFi movies are not in fact documentaries from the future. >>
@willknight Far more problematic is the closing quote, wherein Knight returns to the interviewee he opened with (CEO of a coding tools company) and platforms her opinions about "AI" therapists.
Just so everyone is clear: ChatGPT is still just a language model: just a text synthesis machine/random BS generator. Its training has honed the form of that BA a bit further, including training to avoid things that *look like* certain topics, but there's still no there there.
That "Limitations" section has it wrong though. ChatGPT generates strings based on combinations of words from its training data. When it sometimes appears to say things that are correct and sensible when a human makes sense of them, that's only by chance.
>>
Also the link under "depends on what the model knows" in that screencap points to the "AI Alignment Forum" which looks like one of the message boards from the EA/Longtermist cult. For more on what that is and the damage it's doing, see @timnitgebru 's:
🪜 Building taller and taller ladders won't get you to the moon -- ?
🏃♀️ Running faster doesn't get you closer to teleportation -- me
⏱️ "dramatically improving the precision or efficiency of clock technology does not lead to a time travel device" -- @fchollet
@fchollet All helpful metaphors, I think, for explaining why it's foolish to believe that deep learning (useful as it may be) isn't a path towards what @fchollet calls "cognitive autonomy".
[I couldn't quickly turn up the source for the ladder one, and would be grateful for leads.]
>>
@fchollet Somehow, the current conversation & economy around #AI have left us in a place where the people who claim the opposite don't carry the burden of proof and/or try to discharge it with cherry picked examples.
Facebook (sorry: Meta) AI: Check out our "AI" that lets you access all of humanity's knowledge.
Also Facebook AI: Be careful though, it just makes shit up.
This isn't even "they were so busy asking if they could"—but rather they failed to spend 5 minutes asking if they could.
>>
Using a large LM as a search engine was a bad idea when it was proposed by a search company. It's still a bad idea now, from a social media company. Fortunately, @chirag_shah and I already wrote the paper laying that all out:
Today's #AIhype take-down + analysis (first crossposted to both Twitter & Mastodon): an "AI politician". vice.com/en/article/jgp…
/1
Working from the reporting by @chloexiang at @motherboard, it appears that this is some sort of performance art, except that the project is (purports to be?) interacting with the actual Danish political system.
/2
I have no objections to performance art in general, and something that helps the general public grasp the absurdity of claims of "AI" and reframe what these systems should be used for seems valuable.
/3
I guess it's a milestone for "AI" startups when they get their puff-pieces in the media. I want to highlight some obnoxious things about this one, on Cohere. #AIhype ahead...
First off, it's boring. I wouldn't have made it past the first couple of paragraphs, except the reporter had talked to me so I was (increasingly morbidly) curious how my words were being used.
>>
The second paragraph (and several others) is actually the output of their LLM. This is flagged in the subhead and in the third paragraph. I still think it's terrible journalistic practice.