this post was submitted on 30 Jan 2025
41 points (100.0% liked)
Technology
1060 readers
25 users here now
A tech news sub for communists
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
It's what OpenAI claims https://www.nbcnews.com/tech/tech-news/openai-says-deepseek-may-inapproriately-used-data-rcna189872
Yeah okay so they're literally just talking about using ChatGPT answers in the training set. Everyone does this, it's completely fair game. Hell, the R1 Qwen that everyone is using now to run 'R1' on their own device is Qwen distilled with R1.
(Not that I believe in copyright to begin with).
Exactly and it's just stunning hypocrisy on the part of OpenAI who have been sucking up all data with complete disregard for ownership. Also agree that copyright wouldn't even be necessary in a socialist society, while it only ends up being abused under capitalism.
I would argue that a new socialist country should retain copyright, but drastically shorten its duration (back to the original 7 years or less). The reason being that copyright is the mechanism that protects the copyleft. Without copyright, the GPL would no longer be enforceable. I'm no fan of copyright and would love to see it eventually abolished, but it has its uses today.
Yeah, capitalist relations need to be abolished first.
Yeah the real problem with this is that ChatGPT is going to give a lot of garbage answers when it comes to political questions, and if DeepSeek takes in too much of that, it will also start spitting out very bad takes. It's like copying off of your classmate in a test in school, but that classmate is only known for getting good grades in some classes but not so good grades in others. You need to be careful to only copy on the subjects you know they're actually competent at.
Lol, what's ClosedAI going to do? Hide ChatGPT's answers behind a summary (like with the CoT) 🤣
What is CoT?
Notable OpenAI's O1 cot is hidden with only a summary available sometimes
From Deepseek:
Chain of Thought (CoT) in LLMs refers to a prompting technique that guides large language models to articulate intermediate reasoning steps when solving a problem, mimicking human-like logical progression. Here's a concise breakdown:
Purpose: Enhances performance on complex tasks (e.g., math, logic, commonsense reasoning) by breaking problems into sequential steps, reducing errors from direct, unstructured answers.
Mechanism:
Benefits:
Variants:
Effectiveness: Particularly impactful for tasks requiring structured reasoning, while less critical for simple queries. Research shows marked accuracy gains in benchmarks like math word problems.
In essence, CoT leverages the model's generative capability to externalize reasoning, bridging the gap between opaque model decisions and interpretable human problem-solving.
Example: