Discussion about this post

User's avatar
Alan Goldhammer's avatar

As I was Cosma's next door neighbor when he was growing up, I'll take some credit for what he is doing right now 😊, The best controlled experiment for ChatGPT is for big time booster, Tyler Cowen, to split one of his econ classes in two sections. One section get Professor Cowen and the other ChatGPT as the instructor. A common exam would be given to both groups and results assessed. This is the only way to do it.

vorkosigan1's avatar

From my light interactions with ChatGPT, it's a useful tool if you know what the wrong answers are, and want it to (a) proofread, (b) offer ideas/analysis that you might have come up with, after dozens or hundreds of hours of research, (c) want off the wall ideas to spark your thinking, or (d) want recommendations for purchases. But for ALL of those, if you don't have enough domain knowledge to know when it's in error, it can be worse than useless

For instance, I recently wanted to refer to something as taking "real work". Later in the paragraph, I wanted a callback, so I said something like "The reason this takes real work is...." I was trying to achieve emphasis through repetition, but ChatGPT kept editing it out as making the writing boring. That's a trivial example. I've run into confabulations like those Brad describes, too.

It does occur to me to ask for a Turing-like distinction. Humans confabulate. Humans make mistakes. On an operational level, what is it that distinguishes human thought from GPT LLM? I’m not saying there is no difference; I’m saying I don’t know how to describe it.

6 more comments...

No posts

Ready for more?