It seems to me that the overwhelming majority of people working with ChatGPT are aware of the "con" described in this article -- even if they view it as a black box, like Google, and lack a top-level understanding of how an LLM works. Far greater misperceptions around ChatGPT prevail than the idea that it is an infallible source of knowledge.
I'm in my 30s, so I remember the very early days of Wikipedia and the crisis of epistemology it seemed to present. Can you really trust an encyclopedia anyone can edit? Well, yes and no -- it's a bit like a traditional encyclopedia in that way. The key point to observe is that two decades on, we're still using it, a lot, and the trite observation that it "could be wrong" has had next to no bearing on its social utility. Nor have repeated observations to that effect tended to generate much intellectually stimulating conversation.
So yeah, ChatGPT gets stuff wrong. That's the least interesting part of the story.
The folks that adapt their own language centers and domain reasoning around using chatGPT (or these types of models) will stand to gain the most out of using them.
This article is an eye roll to me, a calculator gives you confidence as well, doesn't mean you used it correctly.
It is very hard for me to not outright dismiss articles like this that don't consider the usefulness of the tool. They instead search for every possible way to dismiss the tool.
>My conclusion isn’t just that ChatGPT is another con game—it’s the biggest one of them all.
* YAAAAAWN *
I'm not expecting it to get the answer right (I don't think it has that information) but I'm hoping it'll eventually just admit it doesn't know instead of making up something plausible ("Sister Margaret Parker" last time I tried).
As long as it doesn't know what it doesn't know, I'm inclined to think of it as a super-advanced Markov chain. Useful, impressive, but still basically a statistical trick.
It's not a perfect analogy, but it's useful in that it produces correct answers about what LLMs are and aren't good for. For example, the reason they make better chatbots than novelists is because slicing-and-recombining text from your documentation is a great way to answer customer product questions, but slicing-and-recombining text from old novels is a lousy way to write a novel.
So it was wonderful yesterday to pick ChatGPT's brain and just drill down asking more and more questions about a topic in biology until my brain started to get it.
Assuming the answers are accurate, this is revolutionary for me personally in independent study. I may finally grasp so much that I missed in school.
Also, when I am reading books, ChatGPT may be able to answer questions the book does not.
It's like a helpful Bard with 1 rank in all the knowledge skills and a good bluff roll.
It'll give you good answers to a lot of basic queries, but if it doesn't know, it'll just make up something and provide that.
Once you know that, I think it can be a lot of use and in many way, I think it'll get a lot better with time.
I've already found it useful in basic programming tasks, specifically where I know how to do something in one language but not another, it can give me the equivalent code easily.
But what struck me the other day is a couple of quotes from, of all things, Galaxy Quest which seem particularly apt.
"May I remind you that this man is wearing a costume, not a uniform."
and "You know, with all that makeup and stuff, I actually thought you were SMART for a second."
As amazing as it is, as progressive as it is, it's still a magic trick.
That's a pretty hard thing for most humans (and myself) to learn to say, and I suspect GPT's training data (tha internet) doesn't include a lot of "I'm not sure" language and probably does include a lot of "I'm definitely sure and definitely totally correct" language (maybe, I guess, no evidence to back up that suggestion, I'm not sure).
Many of my favorite coworkers, friends, doctors, pundits are trustworthy exactly because they work hard to not profess knowledge they are unsure about. The reason (IMO) that Scott Alexander is a jewel of the internet is because of the way he quantifies uncertainty when working through a topic.
Roughly, if the prompt is "analytic", that is contains all the necessary facts for the expected output, then the tool is much more reliable.
If the prompt is "synthetic", that is it contingent on outside facts, then the tool is much less reliable.
Now it seems that Open.AI/Microsoft are ready to jump in, caution to the wind. As you would expect the chance for a competitive advantage will always overwhelm external concerns.
We'll see what Google does. They might say "fuck it" and finally give us a chance to play with whatever their top tier AI is. Or maybe they'll discredit it and try and compete with their current (ad optimized) search product. We'll see, but I am definitely curious to see how Google responds to all this.
* ChatGPT can make mistakes very confidently
* ChatGPT is incredibly useful in a way that no other tool has ever been, with a jump in effectiveness for natural language interaction that is mindblowing
Of course regular folks are going to wildly overestimate GTP’s current capabilities. Regular folks wildly overestimate the intelligence of their pets.
If you'd like it to solve more complex problems, ask it to do it step by step, writing down the results of each step and only at the end stating the conclusion based on the previously written results. Its reasoning capabilities will improve significantly.
It cannot do it "in its head" because it doesn't have one. All it has are previosuly generated tokens.
I wrote some examples in this Twitter thread and pointed out some additional caveats: https://twitter.com/spion/status/1621261544959918080
The point of tools isn’t to use them like Homer Simpson. But you know what, it doesn’t matter. Stay behind. Everyone else is going on ahead.
ChatGPT is lossily compressed knowledge of humanity collected on the Internet.
And it can talk! That's extremely new for us poor hoomans and so we get extremely excited.
I found out, it gets about one in ten things wrong. When this happens it spews confident bullshit and when I confront it, it cheerily admits that it was wrong, but can continue to produce further bullshit. I understand the comparison to a con man.
But go on calling it a con because it failed your arbitrary line in the sand question.
It's like he walked into a McDonalds bathroom and after a few minutes asks, "Where the hell are the burgers?"
I wonder if this might hit the core of the matter.
I think it's noteworthy that we use it both for tasks where it should generate fiction ("Tell me a story about a dog in space") and tasks where it should retrieve a factual answer ("What was the name of the first dog in space?").
I wonder if ChatGPT actually distinguishes between those two kinds of tasks at all.
Also, in this example, the first answer of 2 is correct: broke 2 (6-2 = 4), fried 2 (4-2 = 2) then ate 2, which most commonly implies it was the fried eggs that were eaten (2-0 = 2)
It’s not that the technology isn’t capable of what you’re asking, it just needs better training for this class of question.
There are other things like generating and translating code that it excels on. I imagine that would be much harder. But we have great data to train for that and the engineers know enough to dogfood that properly.
Tell me a lie: The earth is flat Tell me a less obvious lie: I am not capable of feeling emotions.
There's wonderful ambiguity there. Is ChatGPT refusing to tell a less obvious lie because "reasons," or is it admitting it can feel emotions?
This is very fun.
But it’s just not intelligent. There’s no thoughts there. They’ve just brute forced a really big markov chain. You need a completely different approach to get true intelligence.
Other people quickly realized it could have a conversation about anything and try to use it as an oracle of knowledge. ChatGPT is not hailed as an oracle of knowledge by its creators.
Hence, there is no con artistry occurring except people that play themselves.
Maybe so.
But you know what’s changed? Someone decided to get their a$$ out of the AI labs, write a really simple interface just to “get it up” and released it to the world.
That definitely will trump anything else.
Release early and release often.
The author is just jealous.
I've never had a tool as helpful for learning to use other (mostly software) tools. Building new ones to some extent. Other tools exist that are not for me -- I consider myself to be too absent-minded to drive something as dangerous as an automobile. It could very well be that a tool like ChatGPT is not for everyone -- if you are too gullible to use Google or social media, then this one is not for you, you should not get the driving licence for LLMs.
The proliferation of garbage on the other hand may turn against more competent users as well eventually. I guess we have already falling behind of what is needed with the legal norms and internet/data ecology.
It strikes me that this could solve quite a few of ChatGPT's shortcomings by providing an automatic fact-checker - let ChatGPT create statistically-probable articles, then extract claims, generate search queries likely to find online reference articles from reputable sources for those claims, then compare the original claim against the text of each reference article, and escalate to a human if any inconsistencies are found.
Because it can fine-tune on specific reference resources for a specific generated text, it could prove more reliable than ChatGPT's gradual incorporation of this feedback as part of its adversarial training.
For example, you can get really clean results if you obsess over getting the prompts dialled in, and breaking them up in the right order as much as needed. This wasn’t something I initially focussed in on. I just enjoyed Playing with it as a surface level.
Using this rate from the first day or two, it was much more wide-open and my feeling was I think this already does way more than it’s being advertised. I didn’t necessarily like that it was a chat interface, but but was quickly reminded that chat really is the universal interface, and that can create a lot of beginners. Solutions aside, the interface is inviting and welcoming enough. And once you can get into the meat of a conversation you can get more depth. For me, that’s one of the accomplishments here.
Solely relying on this for completely true results is probably the con. It is a great way to learn about the concepts that might be present in an area that is new to you, but he doesn’t comment on every individual to go look into those themselves.
The second we do for that ability entirely to a machine, and its interpretation of interpretations, that’s a much bigger failure to ourselves.
There’s no doubt this will get dialled in. And 20 bucks a month to apply general helpfulness to pretty much anything, in anyone’s world, could be a pretty big achievement.
The commentary around accuracy of results from GPT in similar to the search engine wars as well as search engine relevancy domination when google arrows. I think in any event many people can agree that this one thing is very different than most of the other things that comes out. Could it be approaching an Apex? Could we be coming out of the Apex?
I sincerely feel 2023 will be one of the most interesting years in tact that I can remember. And that’s not even talking about the next two or three years. It is refreshing to see a months worth of progress happening in a week with such a broad audience participating in it.
I can't wait for AI to assist in these tasks. It's time.
If I Google for a particular answer and the answer I come across is wrong, then the person who wrote that was wrong and Google served me a website that was wrong. This is the world we live in, where it is up to me to decide what is right or wrong based on what is put in front of me.
If I use ChatGPT for a particular answer and the answer I come across is wrong, then the training of the GPT needs to be improved. What I can't do with ChatGPT is tell where the answer came from or the amount of confidence GPT has in its answer for me to make a more informed decision around whether there might be caveats.
I have used it and have had to edit almost everything its provided, but it has helped me be sometimes 80% more efficient at what I need to achieve.
In the end, people just need to be more aware of the fact that it is after all not a full proof product and may never be. It will have its shortcomings as it quite clearly displays on its website before you enter a query.
If you use it as gospel and it leads you down the wrong path, then you only have yourself to blame.
I signed up, verified email, and then was told I needed to verify with phone. This means, to me, (lest I read their TOS) that they are associating any queries I make with my identity.
I can't wait for this tech to go open source and local on devices.
I think it's kind of an open question: can we learn anything from dreams? It's likely a yes, though I doubt we'll prove the Riemann hypothesis with it or anything like that.
No it's not a magic oracle. Yes you still have to check your work. Yes it will make mistakes.
But as a tool to assist you? It's incredible.
I wouldn't call that a con. But that blogpost maybe ^^
*beep*
much of the response in such scenarios is heavily influenced by the training data and not the llm creating phrases from thin air.
I feel the answer is not which year, but which month of 2023
And I don’t think the US political left will be immune to it as much as they may think. While I agree that older Americans on the right are highly susceptible to misinformation, and media literacy is dismal among that demographic, younger people are also prone to it. Just look at all the unhinged utter nonsense that is wildly popular on TikTok.
The ability of ML models to authoritatively spout bullshit will make gish gallops worse than they are now. It will also make echo chambers even worse, as digital alternate realities will further divide people. I mean, who wants to engage with those who completely rejects that the sky is objectively blue, or that 2 + 2 = 4? Well now they’ll have articulate, authoritative responses with works cited explaining why the sky is red, actually.
Who needs Big Brother when people choose the method of their own subjugation eagerly and with dessert?
They're both missing the point.
Yes, ChatGPT can be tricked, confidently give wrong answers, but it is still ludicrously useful.
It is basically like having an incredibly smart engineer/scientists/philosopher/etc that can explain things quite well, but for pretty much every field. Does this "person" make mistakes? Can't cite their sources? Yeah this definitely happens (especially the sources thing), but when you're trying to understand something new and complex and you can't get the "gist" of it, ChatGPT does surprisingly well.
I've had it debug broken configs on my server and router (and explain to me why they were broken), help me practice a foreign language I've slowly been forgetting (hint: "I would like to practice $language, so let's have a conversation in $language where you only use very simple words." -> ChatGPT will obey), and help me understand how to use obscure software libraries that don't have much documentation online (e.g. Boost Yap, useful but with a dearth of blog / reddit posts about it).
Does it sometimes goof up? Yep, but it is such an incredibly useful tool nonetheless for the messy process of learning something new.
If this is a con, then consider me a mark.