Cethin,

Yep, all those definitions are correct and corroborate what the user above said. An LLM does not learn like an animal learns. They aren’t intelligent. They only reproduce patterns similar to human speech. These aren’t the same thing. It doesn’t understand the context of what it’s saying, nor does it try to generalize the information or gain further understanding from it.

It may pass the Turing test, but that’s neither a necessary nor sufficient condition for intelligence. It is just a useful metric.

  • All
  • Subscribed
  • Moderated
  • Favorites
  • piracy@lemmy.dbzer0.com
  • localhost
  • All magazines
  • Loading…
    Loading the web debug toolbar…
    Attempt #