World News
A community for discussing events around the World
Rules:
-
Rule 1: posts have the following requirements:
- Post news articles only
- Video links are NOT articles and will be removed.
- Title must match the article headline
- Not United States Internal News
- Recent (Past 30 Days)
- Screenshots/links to other social media sites (Twitter/X/Facebook/Youtube/reddit, etc.) are explicitly forbidden, as are link shorteners.
-
Rule 2: Do not copy the entire article into your post. The key points in 1-2 paragraphs is allowed (even encouraged!), but large segments of articles posted in the body will result in the post being removed. If you have to stop and think "Is this fair use?", it probably isn't. Archive links, especially the ones created on link submission, are absolutely allowed but those that avoid paywalls are not.
-
Rule 3: Opinions articles, or Articles based on misinformation/propaganda may be removed. Sources that have a Low or Very Low factual reporting rating or MBFC Credibility Rating may be removed.
-
Rule 4: Posts or comments that are homophobic, transphobic, racist, sexist, anti-religious, or ableist will be removed. “Ironic” prejudice is just prejudiced.
-
Posts and comments must abide by the lemmy.world terms of service UPDATED AS OF 10/19
-
Rule 5: Keep it civil. It's OK to say the subject of an article is behaving like a (pejorative, pejorative). It's NOT OK to say another USER is (pejorative). Strong language is fine, just not directed at other members. Engage in good-faith and with respect! This includes accusing another user of being a bot or paid actor. Trolling is uncivil and is grounds for removal and/or a community ban.
Similarly, if you see posts along these lines, do not engage. Report them, block them, and live a happier life than they do. We see too many slapfights that boil down to "Mom! He's bugging me!" and "I'm not touching you!" Going forward, slapfights will result in removed comments and temp bans to cool off.
-
Rule 6: Memes, spam, other low effort posting, reposts, misinformation, advocating violence, off-topic, trolling, offensive, regarding the moderators or meta in content may be removed at any time.
-
Rule 7: We didn't USED to need a rule about how many posts one could make in a day, then someone posted NINETEEN articles in a single day. Not comments, FULL ARTICLES. If you're posting more than say, 10 or so, consider going outside and touching grass. We reserve the right to limit over-posting so a single user does not dominate the front page.
We ask that the users report any comment or post that violate the rules, to use critical thinking when reading, posting or commenting. Users that post off-topic spam, advocate violence, have multiple comments or posts removed, weaponize reports or violate the code of conduct will be banned.
All posts and comments will be reviewed on a case-by-case basis. This means that some content that violates the rules may be allowed, while other content that does not violate the rules may be removed. The moderators retain the right to remove any content and ban users.
Lemmy World Partners
News !news@lemmy.world
Politics !politics@lemmy.world
World Politics !globalpolitics@lemmy.world
Recommendations
For Firefox users, there is media bias / propaganda / fact check plugin.
https://addons.mozilla.org/en-US/firefox/addon/media-bias-fact-check/
- Consider including the article’s mediabiasfactcheck.com/ link
view the rest of the comments
OpenAI is so concerned that AI will do x and y bad thing but still pour all these resources into developing it further.
There are other endeavors where a great deal of the effort is put into making it safe. Space travel for example.
I wish that was the case for AI development. AI safety is a notoriously underfunded, understaffed and still overall neglected field.
That concern is feigned, for PR.
OpenAI isn't responsible for what Russians do with it anymore than any company is for how users use their product
If someone knows that what they're about to create is going to do harm like this, they shoulder some of the responsibility for those consequences. They dont just get to wash their hands of it as if they had no idea.
Why not. The people who are to blame are the people commuting the act.
The thing itself has no ethical or moral impact until it's used by a person. I think it feels good to blame an inventor but that's scapegoating the real culprits. Only way I see your argument making sense is if they intended their tools to be user for unethical reasons.
Because people should consider the pros and cons of what they work on not just pretend that none of the responsibility for those cons is theirs. AI is one of the things that could wipe out humanity. Not in the terminator sense but through unparalleled distruption of the economy and by facilitating a wedge between people through the production of propaganda like none that weve ever seen. i.e deepfakes, personally tailored propaganda etc.
Does it? Doesn't that threat exist even without AI. At its current state its a glorified chatbot. Get rid of it, we still have every think tank filled with quants, statisticians, social scientists and marketing teams pushing all that propaganda. Its not AI doing it. Its humans.
But AI does have potential to also develop new medicines. New materials. It has potential for a lot more good.
It also potentially has a lot of potential to give people some powerful pocket access to some basic services they normally wouldn't have. Imagine an AI trained to help people sort out their finances. Act like an r/askdocs. Help with questions about new hobbies.
So where you see panic, other people see hope. And it isn't the inventors job to tell you or others how to use something.
If we destroy ourselves with every bit of advancement then we deserve it. It would be an inevitability.
Yes. Your point?
That AI isn't the issue
I'm saying that the world is facing many threats, and all of them need to be addressed. Including some AI.
What's the threat AI poses exactly?
Well, there are a lot of threats from AI that are popularised today. Some of them are fake, crazy, or stupid, but others are real. Here are some threats that are real, in my opinion.
In trying to use AI for good, we may give too much power to an AI that we do not understand well enough, and whose motivations are not clear. Worse still, this AI might have motivations that we humans could not understand, even if we wanted to. Such an AI may not value humans at all.
Another threat is that a country like China creates an AI and puts it in charge of their soft power program. It may be possible for an AI to be so intelligent that it could manipulate the world in ways that humans physically could not understand. This could allow China to not only literally take over the world, but also ensure that everybody with any kind of power is absolutely thrilled about it. (Anybody without the power to stop this AI would be disregarded in its calculations, no matter how much they hate the AI. This could easily include most humans.)
One thing AI has already been doing for a decade to our great detriment is optimising the ad revenue of search engines and social networks, with a total disregard for all other consequences.
All of these threats are not new. People have been talking about them for years, now. You should pay more attention.
Not to mention that even if one inventor decides not to release their creation, eventually someone else will make something similar.
Would you say then that our efforts to hinder access to dangerous information aren't working?
In that case, would you object to the posting of detailed schematics on the internet for the creation of nuclear weapons?
No I wouldn't. In general I'm oppesed to any hiding/censoring of information
The incentives to continue development are far too great; if one firm abandons the project, that just means that AI will be developed by a less ethical firm. This is why arguing that AI is bad in-and-of-itself is a moderately effective way to reduce the ethics of the average AI developer.