Welcome to DU!
The truly grassroots left-of-center political community where regular people, not algorithms, drive the discussions and set the standards.
Join the community:
Create a free account
Support DU (and get rid of ads!):
Become a Star Member
Latest Breaking News
Editorials & Other Articles
General Discussion
The DU Lounge
All Forums
Issue Forums
Culture Forums
Alliance Forums
Region Forums
Support Forums
Help & Search
General Discussion
In reply to the discussion: College professor had students grade ChatGPT-generated essays. All 63 essays had hallucinated errors [View all]highplainsdem
(58,479 posts)46. See this:
https://spectrum.ieee.org/ai-hallucination
I'm talking about the AI being hyped and used now, despite all the mistakes it makes.
And I find this a silly point to try to make:
The people using AI for business, for example, are not trying to make stuff up. AI they're using may very well do so regardless of their intent.
You wouldn't defend a calculator that often gives incorrect answers by saying that humans can flub math questions, too. I hope you wouldn't, anyway.
For that matter, no one in their right mind would want to use a calculator that didn't work.
It's really pathetic that so many people are eager to use fallible AI whose results have to be checked very carefully.
But I think there are three main reasons for the popularity of ChatGPT and similar LLMs:
1) Gullibility. People expect computer results to be accurate, and they're impressed as well by the chatbot's fluent, authoritative-sounding prose.
2) Laziness. This applies to the cheaters, whether students or adults who think AI can handle chores they don't like or give them the appearance of having skills and talents they don't have - an illusion that will crumble as soon as they're deprived of the AI.
3) Greed. This applies to all the people who think they'll become richer, quickly, using AI, whether those people are employers hoping to lay off employees, or people dreaming of get-rich-quick schemes where AI gives them marketable writing, code, etc.
Whether or not inaccurate outputs can be eliminated through reinforcement learning with human feedback remains to be seen. For now, the usefulness of large language models in generating precise outputs remains limited.
I'm talking about the AI being hyped and used now, despite all the mistakes it makes.
And I find this a silly point to try to make:
Humans make stuff up. The question is will AI's ever make stuff up less than the average person?
The people using AI for business, for example, are not trying to make stuff up. AI they're using may very well do so regardless of their intent.
You wouldn't defend a calculator that often gives incorrect answers by saying that humans can flub math questions, too. I hope you wouldn't, anyway.
For that matter, no one in their right mind would want to use a calculator that didn't work.
It's really pathetic that so many people are eager to use fallible AI whose results have to be checked very carefully.
But I think there are three main reasons for the popularity of ChatGPT and similar LLMs:
1) Gullibility. People expect computer results to be accurate, and they're impressed as well by the chatbot's fluent, authoritative-sounding prose.
2) Laziness. This applies to the cheaters, whether students or adults who think AI can handle chores they don't like or give them the appearance of having skills and talents they don't have - an illusion that will crumble as soon as they're deprived of the AI.
3) Greed. This applies to all the people who think they'll become richer, quickly, using AI, whether those people are employers hoping to lay off employees, or people dreaming of get-rich-quick schemes where AI gives them marketable writing, code, etc.
Edit history
Please sign in to view edit histories.
Recommendations
0 members have recommended this reply (displayed in chronological order):
47 replies
= new reply since forum marked as read
Highlight:
NoneDon't highlight anything
5 newestHighlight 5 most recent replies
RecommendedHighlight replies with 5 or more recommendations

College professor had students grade ChatGPT-generated essays. All 63 essays had hallucinated errors [View all]
highplainsdem
May 2023
OP
" I'm much more worried about the possibility of us reverting to where AI is."
WestMichRad
May 2023
#2
I think the difficulty for most of us is in "context switching". When we're interacting with a tool
erronis
May 2023
#12
The best way to explain the productivity destruction of context switching!
Lucky Luciano
May 2023
#24
Very nice depiction of the tangled mess we weave. I think a lot of current software
erronis
May 2023
#32
Did the Wendy's drive-thru, I handed $20 on an $18.60 order and got $7.40 back. Returned it.
TheBlackAdder
May 2023
#30
The correct change is $1.40. You got $7.40 back. Could the cashier have calculated it on a
progree
May 2023
#31
I think you are right - Garry Kasparov was the Grand Master. And he is now knows a lot about AI.
erronis
May 2023
#13
Not really. Only linked pages. There's a lot of content that isn't accessed without
erronis
May 2023
#14
The future of Chat bots won't be trained on available datasets like the internet.
Yavin4
May 2023
#19
It probably will get better, but it could also stall out on improving accuracy...
Silent3
May 2023
#35