Menu
News
All News
Dungeons & Dragons
Level Up: Advanced 5th Edition
Pathfinder
Starfinder
Warhammer
2d20 System
Year Zero Engine
Industry News
Reviews
Dragon Reflections
White Dwarf Reflections
Columns
Weekly Digests
Weekly News Digest
Freebies, Sales & Bundles
RPG Print News
RPG Crowdfunding News
Game Content
ENterplanetary DimENsions
Mythological Figures
Opinion
Worlds of Design
Peregrine's Nest
RPG Evolution
Other Columns
From the Freelancing Frontline
Monster ENcyclopedia
WotC/TSR Alumni Look Back
4 Hours w/RSD (Ryan Dancey)
The Road to 3E (Jonathan Tweet)
Greenwood's Realms (Ed Greenwood)
Drawmij's TSR (Jim Ward)
Community
Forums & Topics
Forum List
Latest Posts
Forum list
*Dungeons & Dragons
Level Up: Advanced 5th Edition
D&D Older Editions
*TTRPGs General
*Pathfinder & Starfinder
EN Publishing
*Geek Talk & Media
Search forums
Chat/Discord
Resources
Wiki
Pages
Latest activity
Media
New media
New comments
Search media
Downloads
Latest reviews
Search resources
EN Publishing
Store
EN5ider
Adventures in ZEITGEIST
Awfully Cheerful Engine
What's OLD is NEW
Judge Dredd & The Worlds Of 2000AD
War of the Burning Sky
Level Up: Advanced 5E
Events & Releases
Upcoming Events
Private Events
Featured Events
Socials!
EN Publishing
Twitter
BlueSky
Facebook
Instagram
EN World
BlueSky
YouTube
Facebook
Twitter
Twitch
Podcast
Features
Top 5 RPGs Compiled Charts 2004-Present
Adventure Game Industry Market Research Summary (RPGs) V1.0
Ryan Dancey: Acquiring TSR
Q&A With Gary Gygax
D&D Rules FAQs
TSR, WotC, & Paizo: A Comparative History
D&D Pronunciation Guide
Million Dollar TTRPG Kickstarters
Tabletop RPG Podcast Hall of Fame
Eric Noah's Unofficial D&D 3rd Edition News
D&D in the Mainstream
D&D & RPG History
About Morrus
Log in
Register
What's new
Search
Search
Search titles only
By:
Forums & Topics
Forum List
Latest Posts
Forum list
*Dungeons & Dragons
Level Up: Advanced 5th Edition
D&D Older Editions
*TTRPGs General
*Pathfinder & Starfinder
EN Publishing
*Geek Talk & Media
Search forums
Chat/Discord
Menu
Log in
Register
Install the app
Install
Community
General Tabletop Discussion
*Dungeons & Dragons
D&D AI Fail
JavaScript is disabled. For a better experience, please enable JavaScript in your browser before proceeding.
You are using an out of date browser. It may not display this or other websites correctly.
You should upgrade or use an
alternative browser
.
Reply to thread
Message
<blockquote data-quote="EzekielRaiden" data-source="post: 9323771" data-attributes="member: 6790260"><p>I think we're talking past each other.</p><p></p><p>In the context of <em>the kinds of data being processed</em>, semantic content, the "meanings" of things, is where things like evidence, credence, warrant, etc. lie. Unless you have the ability to look at what something <em>means</em>, and not just what its <em>sequence</em> is, you cannot even begin to consider "evidence" and "warrant" etc. If a human were restricted to <em>exclusively</em> making arguments based on the sequence in which (parts of) words appear, without ever being able to ask what any part actually <em>means</em>, they would never be able to even begin talking about whether one word-part (or set thereof) is warranted or not. To speak of warrant, you must know what the words mean; computers do not know that. They only "know" likely and unlikely sequences in which (parts of) words get used by people.</p><p></p><p>As a general rule, the way language-related* AIs work right now "tokenizes" (breaks up and indexes) the words humans use in some given language (such as English) into subword chunks, "tokens," which can then be combined. Using full words is too ponderous, since many sub-word parts (like "ing" and "tion") show up in bazillions of words, while using individual characters fails to capture enough useful structure in how the language is used by people. Subword tokenization is the norm today, mostly because it's significantly more scalable, and LLMs like GPT derive most of their value from having been scaled up to a large size.</p><p></p><p>The token-space for GPT-3, which is an older model since superseded by GPT-4, is somewhere above 12000 tokens, I don't remember the precise value. That means that GPT-3 has crunched through untold billions of lines of human-written text, identified 12000+ word-bits that get used often enough to be worth noting, and indexed them in a set of enormous matrices (really, sets of matrices of a few different sizes, because it turns out you can save a lot of space by doing certain calculations with smaller matrices). Whenever GPT runs--for example, when you give ChatGPT a prompt and ask it to respond--it takes your prompt input, breaks it up into the tokens it recognizes, and then passes the resulting 12000-dimensional vector through dozens of layers of matrix multiplication, with each layer filled with thousands of training-tweaked weights on the matrix multiplication.</p><p></p><p>These weights capture syntactic relationships between the tokens--they encode the ways that those tokens were used in actual words written by actual humans, the sequences and relationships between sequences. That "training" mostly worked by having the AI guess the next word in <em>already-written</em> text, usually failing, and having its numbers altered until it "correctly" "predicts" the next word actually used. Do this trillions of times, and you build up a mountain of matrix multiplication layers that can encode relatively long, relatively far-reaching syntactic relationships. And some of these can be quite impressive, like how GPT-2 invented a fictitious professor from the University of La Paz for that "unicorns in the Andes who could speak perfect English" prompt.</p><p></p><p>But there's a limit to how much the thing can see. It can only "see" a list of tokens up to the size of its input vector (which, as stated, is lover 12000 tokens). That's a huge number of tokens....and nowhere near enough to cover even a novella, let alone a truly long-form work like a textbook. Once you grow beyond that line, GPT and all models like it will start to "lose the plot," because they do not, <em>cannot</em>, actually understand what words <em>mean.</em> They can only "understand" (encode) sequences of words.</p><p></p><p>If you can't understand what even one single word <em>means</em>, the only portion of truth-value you can ever reach is logical validity, since an argument is valid purely on the basis of its syntactic structure, whether it has the right form, regardless of its semantic content, whether it has the right meaning.</p><p></p><p>*Some, trained for translation, tokenize for multiple languages. Others that are image-related tokenize for patterns of pixels. So on.</p></blockquote><p></p>
[QUOTE="EzekielRaiden, post: 9323771, member: 6790260"] I think we're talking past each other. In the context of [I]the kinds of data being processed[/I], semantic content, the "meanings" of things, is where things like evidence, credence, warrant, etc. lie. Unless you have the ability to look at what something [I]means[/I], and not just what its [I]sequence[/I] is, you cannot even begin to consider "evidence" and "warrant" etc. If a human were restricted to [I]exclusively[/I] making arguments based on the sequence in which (parts of) words appear, without ever being able to ask what any part actually [I]means[/I], they would never be able to even begin talking about whether one word-part (or set thereof) is warranted or not. To speak of warrant, you must know what the words mean; computers do not know that. They only "know" likely and unlikely sequences in which (parts of) words get used by people. As a general rule, the way language-related* AIs work right now "tokenizes" (breaks up and indexes) the words humans use in some given language (such as English) into subword chunks, "tokens," which can then be combined. Using full words is too ponderous, since many sub-word parts (like "ing" and "tion") show up in bazillions of words, while using individual characters fails to capture enough useful structure in how the language is used by people. Subword tokenization is the norm today, mostly because it's significantly more scalable, and LLMs like GPT derive most of their value from having been scaled up to a large size. The token-space for GPT-3, which is an older model since superseded by GPT-4, is somewhere above 12000 tokens, I don't remember the precise value. That means that GPT-3 has crunched through untold billions of lines of human-written text, identified 12000+ word-bits that get used often enough to be worth noting, and indexed them in a set of enormous matrices (really, sets of matrices of a few different sizes, because it turns out you can save a lot of space by doing certain calculations with smaller matrices). Whenever GPT runs--for example, when you give ChatGPT a prompt and ask it to respond--it takes your prompt input, breaks it up into the tokens it recognizes, and then passes the resulting 12000-dimensional vector through dozens of layers of matrix multiplication, with each layer filled with thousands of training-tweaked weights on the matrix multiplication. These weights capture syntactic relationships between the tokens--they encode the ways that those tokens were used in actual words written by actual humans, the sequences and relationships between sequences. That "training" mostly worked by having the AI guess the next word in [I]already-written[/I] text, usually failing, and having its numbers altered until it "correctly" "predicts" the next word actually used. Do this trillions of times, and you build up a mountain of matrix multiplication layers that can encode relatively long, relatively far-reaching syntactic relationships. And some of these can be quite impressive, like how GPT-2 invented a fictitious professor from the University of La Paz for that "unicorns in the Andes who could speak perfect English" prompt. But there's a limit to how much the thing can see. It can only "see" a list of tokens up to the size of its input vector (which, as stated, is lover 12000 tokens). That's a huge number of tokens....and nowhere near enough to cover even a novella, let alone a truly long-form work like a textbook. Once you grow beyond that line, GPT and all models like it will start to "lose the plot," because they do not, [I]cannot[/I], actually understand what words [I]mean.[/I] They can only "understand" (encode) sequences of words. If you can't understand what even one single word [I]means[/I], the only portion of truth-value you can ever reach is logical validity, since an argument is valid purely on the basis of its syntactic structure, whether it has the right form, regardless of its semantic content, whether it has the right meaning. *Some, trained for translation, tokenize for multiple languages. Others that are image-related tokenize for patterns of pixels. So on. [/QUOTE]
Insert quotes…
Verification
Post reply
Community
General Tabletop Discussion
*Dungeons & Dragons
D&D AI Fail
Top