Menu
News
All News
Dungeons & Dragons
Level Up: Advanced 5th Edition
Pathfinder
Starfinder
Warhammer
2d20 System
Year Zero Engine
Industry News
Reviews
Dragon Reflections
White Dwarf Reflections
Columns
Weekly Digests
Weekly News Digest
Freebies, Sales & Bundles
RPG Print News
RPG Crowdfunding News
Game Content
ENterplanetary DimENsions
Mythological Figures
Opinion
Worlds of Design
Peregrine's Nest
RPG Evolution
Other Columns
From the Freelancing Frontline
Monster ENcyclopedia
WotC/TSR Alumni Look Back
4 Hours w/RSD (Ryan Dancey)
The Road to 3E (Jonathan Tweet)
Greenwood's Realms (Ed Greenwood)
Drawmij's TSR (Jim Ward)
Community
Forums & Topics
Forum List
Latest Posts
Forum list
*Dungeons & Dragons
Level Up: Advanced 5th Edition
D&D Older Editions
*TTRPGs General
*Pathfinder & Starfinder
EN Publishing
*Geek Talk & Media
Search forums
Chat/Discord
Resources
Wiki
Pages
Latest activity
Media
New media
New comments
Search media
Downloads
Latest reviews
Search resources
EN Publishing
Store
EN5ider
Adventures in ZEITGEIST
Awfully Cheerful Engine
What's OLD is NEW
Judge Dredd & The Worlds Of 2000AD
War of the Burning Sky
Level Up: Advanced 5E
Events & Releases
Upcoming Events
Private Events
Featured Events
Socials!
EN Publishing
Twitter
BlueSky
Facebook
Instagram
EN World
BlueSky
YouTube
Facebook
Twitter
Twitch
Podcast
Features
Top 5 RPGs Compiled Charts 2004-Present
Adventure Game Industry Market Research Summary (RPGs) V1.0
Ryan Dancey: Acquiring TSR
Q&A With Gary Gygax
D&D Rules FAQs
TSR, WotC, & Paizo: A Comparative History
D&D Pronunciation Guide
Million Dollar TTRPG Kickstarters
Tabletop RPG Podcast Hall of Fame
Eric Noah's Unofficial D&D 3rd Edition News
D&D in the Mainstream
D&D & RPG History
About Morrus
Log in
Register
What's new
Search
Search
Search titles only
By:
Forums & Topics
Forum List
Latest Posts
Forum list
*Dungeons & Dragons
Level Up: Advanced 5th Edition
D&D Older Editions
*TTRPGs General
*Pathfinder & Starfinder
EN Publishing
*Geek Talk & Media
Search forums
Chat/Discord
Menu
Log in
Register
Install the app
Install
Community
General Tabletop Discussion
*Geek Talk & Media
How Generative AI's work
JavaScript is disabled. For a better experience, please enable JavaScript in your browser before proceeding.
You are using an out of date browser. It may not display this or other websites correctly.
You should upgrade or use an
alternative browser
.
Reply to thread
Message
<blockquote data-quote="Gorgon Zee" data-source="post: 9290237" data-attributes="member: 75787"><p>Generative AI has been much in the news both in RPG-land and in the general media. I do a fair amount of work in this area, so I thought it might be helpful to give a worked example of how generative AI works. For text, the AI repeatedly tries to guess the next word in the sequence and keeps going until the special "|end|' token is its best guess, which means that it is done.</p><p>It has no idea of truth, falsehood, safety, appropriateness or tone. It just wants to give you a plausible next word; any qualities of that next word are determined by how it was trained and the previous input. This is why, if you start being aggressive with an AI, or give it weird prompts, it gives aggressive or weird answers -- because that output fits the input you have given it best.</p><p></p><p><em>However</em>: Commercial AI's have a "secret sauce" filter on top of the basic operation which modifies this basic algorithm to make it more sane. I'm not covering that layer in this post.</p><p></p><p>So, let's ask an AI: <strong>What is the next item in this sequence: The Sex Pistols, Rolling Stones, Kate Bush?</strong></p><p></p><p>You and I might look at that list and go "huh, British Pop/Rock artists" and that would generate a list of possibilities and we'd select one as an answer. This is not how GenAI works even slightly. Instead it applies its (currently 7-70 billion or so) parameters to the words "What is the next item in this sequence: The Sex Pistols, Rolling Stones, Kate Bush?" and comes up with a set of possibilities for the next word. Actually, not really a word, but a token, which can be part of a word, as we will see. </p><p></p><p>Things to note:</p><ul> <li data-xf-list-type="ul">That is the complete description of what it does: Choose a plausible next token from a sequence of input tokens</li> <li data-xf-list-type="ul">From the set of possible next tokens, it chooses one at random with probabilities proportional to how likely it thinks each token is. </li> <li data-xf-list-type="ul">You can control how random that choice is with the <em>temperature</em> parameter. At zero, it will always choose the most likely answer (and so become mostly deterministic). At values over 1 it gets a little wild ...</li> <li data-xf-list-type="ul">For each output token, it has to run a repeated complex model involving 70 billion parameters. This is not cheap.</li> </ul><p></p><p>-------</p><p></p><p>So, onto our worked example: <strong>What is the next item in this sequence: The Sex Pistols, Rolling Stones, Kate Bush?</strong></p><p></p><p>I'll set the temperature to 1.2 to get some extreme suggestions and ask for 5 different responses. For each response, I show the token chosen and its probability together with the top 5 alternatives that were given. Note that the possible responses all start with the same basic probabilities: <strong>David (27.11%) • The (19.12%) • Mad (10.09%) • No (9.73%) • Led (5.01%)</strong>. That's because they all start with the same token sequence. Once the first token is chosen, then they diverge.</p><p></p><p><strong>Response 1: Not enough information given.</strong></p><p style="margin-left: 20px"><strong>Not</strong> (2.29%) -- <em>David (27.11%) • The (19.12%) • Mad (10.09%) • No (9.73%) • Led (5.01%)</em></p> <p style="margin-left: 20px"><strong>enough</strong> (16.03%) -- <em>possible (44.07%) • a (22.79%) • enough (16.03%) • able (2.26%) • sure (1.65%)</em></p> <p style="margin-left: 20px"><strong>information</strong> (88.32%) -- <em>information (88.32%) • context (6.28%) • data (3.24%) • info (1.63%) • details (0.18%)</em></p> <p style="margin-left: 20px"><strong>given</strong> (80.25%) -- <em>given (80.25%) • provided (17.45%) • to (1.12%) • available (0.85%) • for (0.21%)</em></p> <p style="margin-left: 20px"><strong>.</strong> (93.24%) -- <em>. (93.24%) • <|end|> (6.30%) • to (0.16%) • yet (0.07%) • here (0.06%)</em></p><p></p><p><strong>Response 2: Led Zeppelin</strong></p><p style="margin-left: 20px"><strong>Led</strong> (5.01%) -- <em>David (27.11%) • The (19.12%) • Mad (10.09%) • No (9.73%) • Led (5.01%)</em></p> <p style="margin-left: 20px"><strong>Ze</strong> (99.99%) -- <em>Ze (99.99%) • Z (0.01%) • ze (0.00%) • Ze (0.00%) • -Z (0.00%)</em></p> <p style="margin-left: 20px"><strong>ppelin</strong> (100.00%) -- <em>ppelin (100.00%) • pp (0.00%) • pl (0.00%) • ep (0.00%) • ppe (0.00%)</em></p><p></p><p><strong>Response 3: The Beatles</strong></p><p style="margin-left: 20px"><strong>The</strong> (19.12%) -- <em>David (27.11%) • The (19.12%) • Mad (10.09%) • No (9.73%) • Led (5.01%)</em></p> <p style="margin-left: 20px"><strong>Beatles</strong> (61.99%) -- <em>Beatles (61.99%) • Clash (24.43%) • Smith (3.23%) • Cure (3.03%) • Who (2.50%)</em></p><p></p><p><strong>Response 4: The Velvet Underground</strong></p><p style="margin-left: 20px"><strong>The</strong> (19.12%) -- <em>David (27.11%) • The (19.12%) • Mad (10.09%) • No (9.73%) • Led (5.01%)</em></p> <p style="margin-left: 20px"><strong>Velvet</strong> (2.11%) -- <em>Beatles (61.99%) • Clash (24.43%) • Smith (3.23%) • Cure (3.03%) • Who (2.50%)</em></p> <p style="margin-left: 20px"><strong>Underground</strong> (99.99%) -- <em>Underground (99.99%) • Under (0.01%) • underground (0.00%) • Und (0.00%) • (0.00%)</em></p><p></p><p><strong>Response 5: David Bowie</strong></p><p style="margin-left: 20px"><strong>David</strong> (27.11%) -- <em>David (27.11%) • The (19.12%) • Mad (10.09%) • No (9.73%) • Led (5.01%)</em></p> <p style="margin-left: 20px"><strong>Bowie</strong> (100.00%) -- <em>Bowie (100.00%) • Byrne (0.00%) • Bow (0.00%) • bow (0.00%) • (0.00%)</em></p> <p style="margin-left: 20px"></p><p>Response 5 is actually the most likely -- 'David' has a 27% chance of being chosen (the highest) and then it's almost certain we'll go with "Bowie" rather than "Byrne" or others.</p><p>One other thing to note is that when the AI chooses "The" as the next token, it has no idea at that point what comes next. It's only AFTER we've added 'The' to our token sequence, making the new input sequence "What is the next item in this sequence: The Sex Pistols, Rolling Stones, Kate Bush? The" that it comes up with "Beatles (61.99%) • Clash (24.43%) • Smith (3.23%) • Cure (3.03%) • Who (2.50%)" and (for response 3) chooses "Beatles" or (for response 4) chooses "Velvet" -- and that last one has a really low probability. If we lowered the temperature, we'd be hugely unlikely to see that chosen.</p><p></p><p>---------</p><p></p><p>So, not necessarily RPG-focused, but I hope this post helps understand a little about how this new tech works, how you can use it, and why it does weird things. If you want to post in this thread, please keep it focused on the technology rather than the ethical usage of AI. That is a hugely important discussion, but I'm hoping to keep this thread more focused on how it works, rather than how we should manage it.</p></blockquote><p></p>
[QUOTE="Gorgon Zee, post: 9290237, member: 75787"] Generative AI has been much in the news both in RPG-land and in the general media. I do a fair amount of work in this area, so I thought it might be helpful to give a worked example of how generative AI works. For text, the AI repeatedly tries to guess the next word in the sequence and keeps going until the special "|end|' token is its best guess, which means that it is done. It has no idea of truth, falsehood, safety, appropriateness or tone. It just wants to give you a plausible next word; any qualities of that next word are determined by how it was trained and the previous input. This is why, if you start being aggressive with an AI, or give it weird prompts, it gives aggressive or weird answers -- because that output fits the input you have given it best. [I]However[/I]: Commercial AI's have a "secret sauce" filter on top of the basic operation which modifies this basic algorithm to make it more sane. I'm not covering that layer in this post. So, let's ask an AI: [B]What is the next item in this sequence: The Sex Pistols, Rolling Stones, Kate Bush?[/B] You and I might look at that list and go "huh, British Pop/Rock artists" and that would generate a list of possibilities and we'd select one as an answer. This is not how GenAI works even slightly. Instead it applies its (currently 7-70 billion or so) parameters to the words "What is the next item in this sequence: The Sex Pistols, Rolling Stones, Kate Bush?" and comes up with a set of possibilities for the next word. Actually, not really a word, but a token, which can be part of a word, as we will see. Things to note: [LIST] [*]That is the complete description of what it does: Choose a plausible next token from a sequence of input tokens [*]From the set of possible next tokens, it chooses one at random with probabilities proportional to how likely it thinks each token is. [*]You can control how random that choice is with the [I]temperature[/I] parameter. At zero, it will always choose the most likely answer (and so become mostly deterministic). At values over 1 it gets a little wild ... [*]For each output token, it has to run a repeated complex model involving 70 billion parameters. This is not cheap. [/LIST] ------- So, onto our worked example: [B]What is the next item in this sequence: The Sex Pistols, Rolling Stones, Kate Bush?[/B] I'll set the temperature to 1.2 to get some extreme suggestions and ask for 5 different responses. For each response, I show the token chosen and its probability together with the top 5 alternatives that were given. Note that the possible responses all start with the same basic probabilities: [B]David (27.11%) • The (19.12%) • Mad (10.09%) • No (9.73%) • Led (5.01%)[/B]. That's because they all start with the same token sequence. Once the first token is chosen, then they diverge. [B]Response 1: Not enough information given.[/B] [INDENT][B]Not[/B] (2.29%) -- [I]David (27.11%) • The (19.12%) • Mad (10.09%) • No (9.73%) • Led (5.01%)[/I][/INDENT] [INDENT][B]enough[/B] (16.03%) -- [I]possible (44.07%) • a (22.79%) • enough (16.03%) • able (2.26%) • sure (1.65%)[/I][/INDENT] [INDENT][B]information[/B] (88.32%) -- [I]information (88.32%) • context (6.28%) • data (3.24%) • info (1.63%) • details (0.18%)[/I][/INDENT] [INDENT][B]given[/B] (80.25%) -- [I]given (80.25%) • provided (17.45%) • to (1.12%) • available (0.85%) • for (0.21%)[/I][/INDENT] [INDENT][B].[/B] (93.24%) -- [I]. (93.24%) • <|end|> (6.30%) • to (0.16%) • yet (0.07%) • here (0.06%)[/I][/INDENT] [B]Response 2: Led Zeppelin[/B] [INDENT][B]Led[/B] (5.01%) -- [I]David (27.11%) • The (19.12%) • Mad (10.09%) • No (9.73%) • Led (5.01%)[/I][/INDENT] [INDENT][B]Ze[/B] (99.99%) -- [I]Ze (99.99%) • Z (0.01%) • ze (0.00%) • Ze (0.00%) • -Z (0.00%)[/I][/INDENT] [INDENT][B]ppelin[/B] (100.00%) -- [I]ppelin (100.00%) • pp (0.00%) • pl (0.00%) • ep (0.00%) • ppe (0.00%)[/I][/INDENT] [B]Response 3: The Beatles[/B] [INDENT][B]The[/B] (19.12%) -- [I]David (27.11%) • The (19.12%) • Mad (10.09%) • No (9.73%) • Led (5.01%)[/I][/INDENT] [INDENT][B]Beatles[/B] (61.99%) -- [I]Beatles (61.99%) • Clash (24.43%) • Smith (3.23%) • Cure (3.03%) • Who (2.50%)[/I][/INDENT] [B]Response 4: The Velvet Underground[/B] [INDENT][B]The[/B] (19.12%) -- [I]David (27.11%) • The (19.12%) • Mad (10.09%) • No (9.73%) • Led (5.01%)[/I][/INDENT] [INDENT][B]Velvet[/B] (2.11%) -- [I]Beatles (61.99%) • Clash (24.43%) • Smith (3.23%) • Cure (3.03%) • Who (2.50%)[/I][/INDENT] [INDENT][B]Underground[/B] (99.99%) -- [I]Underground (99.99%) • Under (0.01%) • underground (0.00%) • Und (0.00%) • (0.00%)[/I][/INDENT] [B]Response 5: David Bowie[/B] [INDENT][B]David[/B] (27.11%) -- [I]David (27.11%) • The (19.12%) • Mad (10.09%) • No (9.73%) • Led (5.01%)[/I][/INDENT] [INDENT][B]Bowie[/B] (100.00%) -- [I]Bowie (100.00%) • Byrne (0.00%) • Bow (0.00%) • bow (0.00%) • (0.00%)[/I][/INDENT] [INDENT][/INDENT] Response 5 is actually the most likely -- 'David' has a 27% chance of being chosen (the highest) and then it's almost certain we'll go with "Bowie" rather than "Byrne" or others. One other thing to note is that when the AI chooses "The" as the next token, it has no idea at that point what comes next. It's only AFTER we've added 'The' to our token sequence, making the new input sequence "What is the next item in this sequence: The Sex Pistols, Rolling Stones, Kate Bush? The" that it comes up with "Beatles (61.99%) • Clash (24.43%) • Smith (3.23%) • Cure (3.03%) • Who (2.50%)" and (for response 3) chooses "Beatles" or (for response 4) chooses "Velvet" -- and that last one has a really low probability. If we lowered the temperature, we'd be hugely unlikely to see that chosen. --------- So, not necessarily RPG-focused, but I hope this post helps understand a little about how this new tech works, how you can use it, and why it does weird things. If you want to post in this thread, please keep it focused on the technology rather than the ethical usage of AI. That is a hugely important discussion, but I'm hoping to keep this thread more focused on how it works, rather than how we should manage it. [/QUOTE]
Insert quotes…
Verification
Post reply
Community
General Tabletop Discussion
*Geek Talk & Media
How Generative AI's work
Top