194
submitted 5 months ago by alb_004@lemm.ee to c/technology@lemmy.world
you are viewing a single comment's thread
view the rest of the comments
[-] givesomefucks@lemmy.world 23 points 5 months ago

If scientists made AI, then it wouldn't be an issue for AI to say "I don't know".

But capitalists are making it, and the last thing you want is it to tell an investor "I don't know". So you tell it to make up bullshit instead, and hope the investor believes it.

It's a terrible fucking way to go about things, but this is America...

[-] VeganCheesecake@lemmy.blahaj.zone 28 points 5 months ago

Uh, I understand the sentiment, but the model doesn't know anything. And it's legit really hard to differentiate between factual things and random bullshit it made up.

[-] DudeDudenson@lemmings.world 18 points 5 months ago

Was gonna say, the AI doesn't make up or admit bullshit, its just a very advanced a prediction algorithm. It responds with what the combination of words that is most likely the expected answer.

Wether that is accurate or not is part of training it but you'll never get 100% accuracy to any query

[-] maynarkh@feddit.nl 1 points 5 months ago

If it can name what the most likely combination is, couldn't it also know how likely that combination of words is?

[-] DudeDudenson@lemmings.world 7 points 5 months ago* (last edited 5 months ago)

It's not actually deciding anything, the AI thinking is marketing fluff really. But yes that's called confidence rating and it does. But at the scale of something like chatgpt that uses a snapshot of the entire internet and is non mutable there's no way to train it for every possible question. If you ask about a topic 99% of the internet gets wrong it'll respond the wrong thing with 99% confidence

[-] wahming@monyet.cc 3 points 5 months ago

No, because that requires it to understand the words. It doesn't.

[-] kent_eh@lemmy.ca 3 points 5 months ago

If it has been trained using questionable sources, or if it's training data includes sarcastic responses (without understanding that context), it isn't hard to imagine how confidently wrong some of the responses could be.

[-] Bishma@discuss.tchncs.de 8 points 5 months ago

Yeah, no one can make it say "I don't know" because it is not really AI. Business bros decided to call it that and everyone smiled and nodded. LLMs are 1 small component (maybe) of AI. Maybe 1/80th of a true AI or AGI.

Honestly the most impressive part of LLMs is the tokenizer that breaks down the request, not the predictive text button masher that comes up with the response.

[-] Kichae@lemmy.ca 10 points 5 months ago

Honestly the most impressive part of LLMs is the tokenizer that breaks down the request, not the predictive text button masher that comes up with the response.

Yes, exactly! It's ability to parse the input is incredible. It's the thing that has that "wow" factor, and it feels downright magical.

Unfortunately, that also makes people intuitively trust its output.

load more comments (35 replies)
this post was submitted on 29 Apr 2024
194 points (94.9% liked)

Technology

58737 readers
4426 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS