this post was submitted on 13 Jun 2023
32 points (100.0% liked)

Technology

37602 readers
389 users here now

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 2 years ago
MODERATORS
 

We are keeping a list of AI/ML related links from research to more accessible items, hoping to share some of the more accessible posts with a wider community!

you are viewing a single comment's thread
view the rest of the comments
[–] manitcor@lemmy.intai.tech 4 points 1 year ago* (last edited 1 year ago)

Not wrong, but not right, it does not "know" and is not fundamentally capable of "analysis" they way we describe it. Its a weighted tokenmap, what makes it special is how those weights were derived. There are papers on self-reflection that are essentially processes in mapping those weights. You can assess the probabilities of your token response in the API as well.

What's interesting is that sufficiently large models demonstrate the ability to build upon themselves using similar techniques applied in training and in fine tuning right in the prompt. This means you can use a combination of reflective conversation and embedding to create prompts that act like fine-tuned agents. Great for fast prototyping and cheaper than a fine tune run!