this post was submitted on 25 May 2024
775 points (97.1% liked)

Technology

59422 readers
2897 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

The research from Purdue University, first spotted by news outlet Futurism, was presented earlier this month at the Computer-Human Interaction Conference in Hawaii and looked at 517 programming questions on Stack Overflow that were then fed to ChatGPT.

“Our analysis shows that 52% of ChatGPT answers contain incorrect information and 77% are verbose,” the new study explained. “Nonetheless, our user study participants still preferred ChatGPT answers 35% of the time due to their comprehensiveness and well-articulated language style.”

Disturbingly, programmers in the study didn’t always catch the mistakes being produced by the AI chatbot.

“However, they also overlooked the misinformation in the ChatGPT answers 39% of the time,” according to the study. “This implies the need to counter misinformation in ChatGPT answers to programming questions and raise awareness of the risks associated with seemingly correct answers.”

you are viewing a single comment's thread
view the rest of the comments
[–] chaosCruiser@futurology.today 2 points 5 months ago (1 children)

Looking forward to it, but won’t be disappointed if it takes a bit longer than expected.

[–] AIhasUse@lemmy.world 1 points 5 months ago (1 children)

Ray Kurzweil has a phenomenal record of making predictions. He's like 90% or something and has been saying AGI by 2029 for something like 30+ years. Last I heard, he is sticking with it, but he admits he may be a year or two off in either direction. AGI is a pretty broad term, but if you take it as "better than nearly every human in every field of expertise," then I think 2029 is quite reasonable.

[–] chaosCruiser@futurology.today 4 points 5 months ago (1 children)

That’s not very far in the future, so it’s going to be really exciting to see how that works out.

[–] explodicle@sh.itjust.works 3 points 5 months ago (1 children)

Maybe only 51% of the code it writes needs to be good before it can self-improve. In which case, we're nearly there!

[–] AIhasUse@lemmy.world 2 points 5 months ago

We are already past that. The 48% is from a version of chatgpt(3.5) that came out a year ago, there has been lots of progress since then.