this post was submitted on 13 Jul 2023
31 points (100.0% liked)

Technology

19 readers
2 users here now

This magazine is dedicated to discussions on the latest developments, trends, and innovations in the world of technology. Whether you are a tech enthusiast, a developer, or simply curious about the latest gadgets and software, this is the place for you. Here you can share your knowledge, ask questions, and engage in discussions on topics such as artificial intelligence, robotics, cloud computing, cybersecurity, and more. From the impact of technology on society to the ethical considerations of new technologies, this category covers a wide range of topics related to technology. Join the conversation and let's explore the ever-evolving world of technology together!

founded 2 years ago
 

I know a lot of people want to interpret copyright law so that allowing a machine to learn concepts from a copyrighted work is copyright infringement, but I think what people will need to consider is that all that's going to do is keep AI out of the hands of regular people and place it specifically in the hands of people and organizations who are wealthy and powerful enough to train it for their own use.

If this isn't actually what you want, then what's your game plan for placing copyright restrictions on AI training that will actually work? Have you considered how it's likely to play out? Are you going to be able to stop Elon Musk, Mark Zuckerberg, and the NSA from training an AI on whatever they want and using it to push propaganda on the public? As far as I can tell, all that copyright restrictions will accomplish to to concentrate the power of AI (which we're only beginning to explore) in the hands of the sorts of people who are the least likely to want to do anything good with it.

I know I'm posting this in a hostile space, and I'm sure a lot of people here disagree with my opinion on how copyright should (and should not) apply to AI training, and that's fine (the jury is literally still out on that). What I'm interested in is what your end game is. How do you expect things to actually work out if you get the laws that you want? I would personally argue that an outcome where Mark Zuckerberg gets AI and the rest of us don't is the absolute worst possibility.

you are viewing a single comment's thread
view the rest of the comments
[–] IncognitoErgoSum@kbin.social 2 points 1 year ago (1 children)

So clearly we do agree on most of this stuff, but I did want to point out a possibility you may not have considered.

If we're just talking about what you can do, then these laws aren't going to matter because you can just pirate whatever training material you want.

This depends on the penalty and how strictly it's enforced. If it's enforced like normal copyright law, then you're right; your chances of getting in serious trouble just for downloading stuff are essentially nil -- the worst thing that will happen to you is your ISP will three-strikes you and you'll lose internet access. On the other hand, there's a lot of panic surrounding AI, and the government might use that as an excuse to pass laws that would give people prison time for possessing one, and then fund strict enforcement. I hope that doesn't happen, but with rumblings of insane laws that would give people prison time for using a VPN to watch a TV show outside of the country, I'm a bit concerned.

As for the parent comment's motivations, it's hard to say for sure with any particular individual, but I have noticed a pattern among neoliberals where they say things like "well, the rich are already powerful and we can't do anything about it, so why try" or "having universal health care, which the rest of the first world has implemented successfully, is unrealistic, so why try" and so on. It often boils down to giving lip service to progressive social values while steadfastly refusing to do anything that might actually make a difference. It's economic conservatism dressed as progressivism. Even if that's not what they meant (and it would be unwise of me to just assume that), I feel like that general attitude needs to be confronted.

[–] FaceDeer@kbin.social 2 points 1 year ago (1 children)

If I'm the "parent comment" you're referring to, then that's very much not my motivation. I'm just pointing out that "AI is accessible to everyone" is not a hard binary situation, and that while it may be true that big giant corporations have an advantage due to being big giant corporations with a ton of resources to throw at this stuff AI is indeed still accessible to some degree to the average consumer.

Well, again, "the average consumer" being first-world individuals with the resources to buy a nice computer and spend time playing with it. These things are a continuum and that's not the end point of it, you can always go further down the resource rankings and find people for whom AI is not "accessible" by whatever standards. Unfortunately it's kind of accepted as a given that people on the poor end of the spectrum don't have access to this kind of stuff or will have to depend on external service providers.

[–] IncognitoErgoSum@kbin.social 1 points 1 year ago (1 children)

If I'm the "parent comment" you're referring to, then that's very much not my motivation.

You're not. I was talking about the thread parent: "Many things in life are a privilege for these groups. AI is no different." I should have been more specific.

At any rate, I personally feel that we have a moral responsibility to make it accessible to as many people as possible.

[–] FaceDeer@kbin.social 3 points 1 year ago

Okay, just wanted to make sure since I'm upstream of the comment.

I agree that making these things as accessible as possible is ideal, it's just that the "as possible" part is tricky with expensive new technology like this. My personal desire is to see UBI implemented on the backs of AI and robot labor, which hopefully will come a lot closer to making universal access possible.