227
submitted 1 year ago by d3Xt3r@beehaw.org to c/technology@beehaw.org

I asked Google Bard whether it thought Web Environment Integrity was a good or bad idea. Surprisingly, not only did it respond that it was a bad idea, it even went on to urge Google to drop the proposal.

you are viewing a single comment's thread
view the rest of the comments
[-] localhost@beehaw.org 44 points 1 year ago* (last edited 1 year ago)

That's not entirely true.

LLMs are trained to predict next word given context, yes. But in order to do that, they develop internal model that minimizes error across wide range of contexts - and emergent feature of this process is that the model DOES perform more than pure compression of the training data.

For example, GPT-3 is able to calculate addition and subtraction problems that didn't appear in the training dataset. This would suggest that the model learned how to perform addition and subtraction, likely because it was easier or more efficient than storing all of the examples from the training data separately.

This is a simple to measure example, but it's enough to suggests that LLMs are able to extrapolate from the training data and perform more than just stitch relevant parts of the dataset together.

[-] fuzzzerd@programming.dev 8 points 1 year ago

That's interesting, I'd be curious to read more about that. Do you have any links to get started with? Searching this type of stuff on Google yields less than ideal results.

[-] localhost@beehaw.org 7 points 1 year ago

In my comment I've been referencing https://arxiv.org/pdf/2005.14165.pdf, specifically section 3.9.1 where they summarize results of the arithmetic tasks.

[-] hikaru755@feddit.de 6 points 1 year ago

Check out this one: https://thegradient.pub/othello/

In it, researchers built a custom LLM trained to play a board game just by predicting the next move in a series of moves, with no input at all about the game state. They found evidence of an internal representation of the current game state, although the model had never been told what that game state looks like.

[-] Xandolas@beehaw.org 2 points 1 year ago

isn't gpt famously bad at math problems?

[-] localhost@beehaw.org 6 points 1 year ago

GPT3 is pretty bad at it compared to alternatives (although it's hard to compete with calculators on that field), but if it was just repeating after the training dataset it would be way worse. From the study I've linked in my other comment (https://arxiv.org/pdf/2005.14165.pdf):

On addition and subtraction, GPT-3 displays strong proficiency when the number of digits is small, achieving 100% accuracy on 2 digit addition, 98.9% at 2 digit subtraction, 80.2% at 3 digit addition, and 94.2% at 3-digit subtraction. Performance decreases as the number of digits increases, but GPT-3 still achieves 25-26% accuracy on four digit operations and 9-10% accuracy on five digit operations, suggesting at least some capacity to generalize to larger numbers of digits.

To spot-check whether the model is simply memorizing specific arithmetic problems, we took the 3-digit arithmetic problems in our test set and searched for them in our training data in both the forms " + =" and " plus ". Out of 2,000 addition problems we found only 17 matches (0.8%) and out of 2,000 subtraction problems we found only 2 matches (0.1%), suggesting that only a trivial fraction of the correct answers could have been memorized. In addition, inspection of incorrect answers reveals that the model often makes mistakes such as not carrying a “1”, suggesting it is actually attempting to perform the relevant computation rather than memorizing a table.

this post was submitted on 13 Aug 2023
227 points (100.0% liked)

Technology

37727 readers
493 users here now

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 2 years ago
MODERATORS