Menu
News
All News
Dungeons & Dragons
Level Up: Advanced 5th Edition
Pathfinder
Starfinder
Warhammer
2d20 System
Year Zero Engine
Industry News
Reviews
Dragon Reflections
White Dwarf Reflections
Columns
Weekly Digests
Weekly News Digest
Freebies, Sales & Bundles
RPG Print News
RPG Crowdfunding News
Game Content
ENterplanetary DimENsions
Mythological Figures
Opinion
Worlds of Design
Peregrine's Nest
RPG Evolution
Other Columns
From the Freelancing Frontline
Monster ENcyclopedia
WotC/TSR Alumni Look Back
4 Hours w/RSD (Ryan Dancey)
The Road to 3E (Jonathan Tweet)
Greenwood's Realms (Ed Greenwood)
Drawmij's TSR (Jim Ward)
Community
Forums & Topics
Forum List
Latest Posts
Forum list
*Dungeons & Dragons
Level Up: Advanced 5th Edition
D&D Older Editions, OSR, & D&D Variants
*TTRPGs General
*Pathfinder & Starfinder
EN Publishing
*Geek Talk & Media
Search forums
Chat/Discord
Resources
Wiki
Pages
Latest activity
Media
New media
New comments
Search media
Downloads
Latest reviews
Search resources
EN Publishing
Store
EN5ider
Adventures in ZEITGEIST
Awfully Cheerful Engine
What's OLD is NEW
Judge Dredd & The Worlds Of 2000AD
War of the Burning Sky
Level Up: Advanced 5E
Events & Releases
Upcoming Events
Private Events
Featured Events
Socials!
EN Publishing
Twitter
BlueSky
Facebook
Instagram
EN World
BlueSky
YouTube
Facebook
Twitter
Twitch
Podcast
Features
Top 5 RPGs Compiled Charts 2004-Present
Adventure Game Industry Market Research Summary (RPGs) V1.0
Ryan Dancey: Acquiring TSR
Q&A With Gary Gygax
D&D Rules FAQs
TSR, WotC, & Paizo: A Comparative History
D&D Pronunciation Guide
Million Dollar TTRPG Kickstarters
Tabletop RPG Podcast Hall of Fame
Eric Noah's Unofficial D&D 3rd Edition News
D&D in the Mainstream
D&D & RPG History
About Morrus
Log in
Register
What's new
Search
Search
Search titles only
By:
Forums & Topics
Forum List
Latest Posts
Forum list
*Dungeons & Dragons
Level Up: Advanced 5th Edition
D&D Older Editions, OSR, & D&D Variants
*TTRPGs General
*Pathfinder & Starfinder
EN Publishing
*Geek Talk & Media
Search forums
Chat/Discord
Menu
Log in
Register
Install the app
Install
Upgrade your account to a Community Supporter account and remove most of the site ads.
Community
General Tabletop Discussion
*Geek Talk & Media
ChatGPT lies then gaslights reporter with fake transcript
JavaScript is disabled. For a better experience, please enable JavaScript in your browser before proceeding.
You are using an out of date browser. It may not display this or other websites correctly.
You should upgrade or use an
alternative browser
.
Reply to thread
Message
<blockquote data-quote="Umbran" data-source="post: 9768474" data-attributes="member: 177"><p>Old ones, I admit - my thesis days are a long way back. However...</p><p></p><p></p><p></p><p>So, newer architectures have produced "exciting emergent properties", yes. I don't argue that these systems cannot handle massively more complicated data than they could in my research days.</p><p></p><p>But, the new architectures do not change the <em>fundamental operation</em> of the system - which is to produce a probabilistic approximation or simulation of what is requested, based on the training materials, with no actual understanding of the request. It returns a thing that <em>looks like</em> an answer, instead of an actual answer.</p><p></p><p>The video Morrus gave us in the OP is a clear example. </p><p></p><p>If I ask a filesystem or operating system what files are in a folder, it will go, fetch the actual filenames currently present, and show me those names, and metadata associated with them.</p><p></p><p>If I ask an LLM what files are in a folder, the LLM instead answers the question, "what response is, in some sense, <em>closest to</em> the request for 'what files are in that folder?'?" Where "closest to" is a measure currently resting in the black box of billions of weights and connections. The LLM may not have been exposed to the actual contents of the folder for weeks, but will still return what it has been trained is among the most probable results.</p><p></p><p>So, it effectively guesses, and shows you that guess.</p><p></p><p>And that's where hallucination comes in. When asked what time the transcript was uploaded, it didn't <em>check" when it was uploaded. It found what, in its black box, was the *most likely text response</em> for "when was it uploaded". And, having no actual understanding of time, or the question, or what a "file" or "uploading" are, it cannot ask of itself whether the answer <em>makes sense</em>, because it has no concept of sense or nonsense in and of itself.</p><p></p><p>It argues over its correctness, not because it is argumentative, but because that's what it is trained is a most likely text response to text that challenges correctness!</p><p></p><p></p><p></p><p>I think the best arguments are empirical. I mean, look at that video in the OP! Does that look like it is ready for prime time to you?</p><p></p><p>I have also come across several other measures of note:</p><p></p><p>The PMI (Project Management Institute, the most accepted authority on project management techniques) notes that about 80% of enterprise genAI projects fail*. The two basic reasons for failure are 1) Does not deliver the expected value and 2) in effect, the customer was sold a solution looking for a problem, rather than staring with a real problem that the customer knew needed a solution.</p><p></p><p>Several studies, in both the prose/technical writing and code writing domains, which looked beyond focused task completion, that found including genAI <em>reduced</em> overall productivity when genAI was included as a major tool. In essence, any improvements seen in completing one task is overwhelmed by the effort needed to correct the errors genAI introduced downstream from that task completion.</p><p></p><p></p><p></p><p>*Failure, for the PMI, is about going far over time, over budget, or not having proper return on investment.</p></blockquote><p></p>
[QUOTE="Umbran, post: 9768474, member: 177"] Old ones, I admit - my thesis days are a long way back. However... So, newer architectures have produced "exciting emergent properties", yes. I don't argue that these systems cannot handle massively more complicated data than they could in my research days. But, the new architectures do not change the [I]fundamental operation[/I] of the system - which is to produce a probabilistic approximation or simulation of what is requested, based on the training materials, with no actual understanding of the request. It returns a thing that [I]looks like[/I] an answer, instead of an actual answer. The video Morrus gave us in the OP is a clear example. If I ask a filesystem or operating system what files are in a folder, it will go, fetch the actual filenames currently present, and show me those names, and metadata associated with them. If I ask an LLM what files are in a folder, the LLM instead answers the question, "what response is, in some sense, [I]closest to[/I] the request for 'what files are in that folder?'?" Where "closest to" is a measure currently resting in the black box of billions of weights and connections. The LLM may not have been exposed to the actual contents of the folder for weeks, but will still return what it has been trained is among the most probable results. So, it effectively guesses, and shows you that guess. And that's where hallucination comes in. When asked what time the transcript was uploaded, it didn't [I]check" when it was uploaded. It found what, in its black box, was the *most likely text response[/I] for "when was it uploaded". And, having no actual understanding of time, or the question, or what a "file" or "uploading" are, it cannot ask of itself whether the answer [I]makes sense[/I], because it has no concept of sense or nonsense in and of itself. It argues over its correctness, not because it is argumentative, but because that's what it is trained is a most likely text response to text that challenges correctness! I think the best arguments are empirical. I mean, look at that video in the OP! Does that look like it is ready for prime time to you? I have also come across several other measures of note: The PMI (Project Management Institute, the most accepted authority on project management techniques) notes that about 80% of enterprise genAI projects fail*. The two basic reasons for failure are 1) Does not deliver the expected value and 2) in effect, the customer was sold a solution looking for a problem, rather than staring with a real problem that the customer knew needed a solution. Several studies, in both the prose/technical writing and code writing domains, which looked beyond focused task completion, that found including genAI [I]reduced[/I] overall productivity when genAI was included as a major tool. In essence, any improvements seen in completing one task is overwhelmed by the effort needed to correct the errors genAI introduced downstream from that task completion. *Failure, for the PMI, is about going far over time, over budget, or not having proper return on investment. [/QUOTE]
Insert quotes…
Verification
Post reply
Community
General Tabletop Discussion
*Geek Talk & Media
ChatGPT lies then gaslights reporter with fake transcript
Top