this post was submitted on 13 Jul 2023
31 points (100.0% liked)

Technology

19 readers
2 users here now

This magazine is dedicated to discussions on the latest developments, trends, and innovations in the world of technology. Whether you are a tech enthusiast, a developer, or simply curious about the latest gadgets and software, this is the place for you. Here you can share your knowledge, ask questions, and engage in discussions on topics such as artificial intelligence, robotics, cloud computing, cybersecurity, and more. From the impact of technology on society to the ethical considerations of new technologies, this category covers a wide range of topics related to technology. Join the conversation and let's explore the ever-evolving world of technology together!

founded 2 years ago
 

I know a lot of people want to interpret copyright law so that allowing a machine to learn concepts from a copyrighted work is copyright infringement, but I think what people will need to consider is that all that's going to do is keep AI out of the hands of regular people and place it specifically in the hands of people and organizations who are wealthy and powerful enough to train it for their own use.

If this isn't actually what you want, then what's your game plan for placing copyright restrictions on AI training that will actually work? Have you considered how it's likely to play out? Are you going to be able to stop Elon Musk, Mark Zuckerberg, and the NSA from training an AI on whatever they want and using it to push propaganda on the public? As far as I can tell, all that copyright restrictions will accomplish to to concentrate the power of AI (which we're only beginning to explore) in the hands of the sorts of people who are the least likely to want to do anything good with it.

I know I'm posting this in a hostile space, and I'm sure a lot of people here disagree with my opinion on how copyright should (and should not) apply to AI training, and that's fine (the jury is literally still out on that). What I'm interested in is what your end game is. How do you expect things to actually work out if you get the laws that you want? I would personally argue that an outcome where Mark Zuckerberg gets AI and the rest of us don't is the absolute worst possibility.

you are viewing a single comment's thread
view the rest of the comments
[–] veridicus@kbin.social 2 points 1 year ago (1 children)

The AIs we're talking about are neural networks. They don't do statistics, they don't have databases, and they don't take mathematical averages. They simulate neurons, and their ability to learn concepts is emergent from that, the same way the human brain is.

This is not at all accurate. Yes, there are very immature neural simulation systems that are being prototyped but that's not what you're seeing in the news today. What the public is witnessing is fundamentally based on vector mathematics. It's pure math and there is nothing at all emergent about it.

If an artist uses a copyrighted work on their mood board or as inspiration, then they should pay for that, because they're making a profit from that copyrighted work.

That's not how copyright works, nor should it. Anyone who creates a mood board from a blank slate is using their learned experience, most of which they gathered from other works. If you were to write a book analyzing movies, for example, you shouldn't have to pay the copyright for all those movies. You can make a YouTube video right now with a few short clips from a movie or quotes from a book and you're not violating copyright. You're just not allowed to make a largely derivative work.

[–] IncognitoErgoSum@kbin.social 3 points 1 year ago* (last edited 1 year ago) (1 children)

So to clarify, are you making the claim that nothing that's simulated with vector mathematics can have emergent properties? And that AIs like GPT and Stable Diffusion don't contain simulated neurons?

[–] veridicus@kbin.social 1 points 1 year ago (1 children)

Yes, and the math is all publicly documented.

[–] IncognitoErgoSum@kbin.social 3 points 1 year ago (1 children)
[–] veridicus@kbin.social 1 points 1 year ago (2 children)

No, I'm not your Google. You can easily read the background of Stable Diffusion and see it's based on Markov chains.

[–] IncognitoErgoSum@kbin.social 1 points 1 year ago (1 children)

LOL, I love kbin's public downvote records. I quoted a bunch of different sources demonstrating that you're wrong, and rather than own up to it and apologize for preaching from atop Mt. Dunning-Kruger, you downvoted me and ran off.

I advise you to step out of whatever echo chamber you've holed yourself up in and learn a bit about AI before opining on it further.

[–] veridicus@kbin.social 1 points 1 year ago* (last edited 1 year ago) (1 children)

My last response didn’t post for some reason. The mistake you’re making is that a neural network is not a neural simulation. It’s relatively simple math, just on a very large scale. I think you mentioned earlier, for example, you played with PyTorch. You should then know that NN stack is based on vector math. You’re making assumptions based on terminology but when you read deeper you’ll see what I mean.

[–] IncognitoErgoSum@kbin.social 1 points 1 year ago

I said it was a neural network.

You said it wasn't.

I asked you for a link.

You told me to do your homework for you.

I did your homework. Your homework says it's a neural network. I suggest you read it, since I took the time to find it for you.

Anyone who knows the first thing about neural networks knows that, yes, artificial neurons are simulated with matrix multiplications, why is why people use GPUs to do them. The simulations are not down to the molecule becuase they don't need to be. The individual neurons are relatively simple math, but when you get into billions of something, you don't need extreme complexity for new properties to emerge (in fact, the whole idea of emergent properties is that they arise from collections of simple things, like the rules of the Game of Life, for instance, which are far simpler than simulated neurons). Nothing about this makes me wrong about what I'm talking about for the purposes of copyright. Neural networks store concepts. They don't archive copies of data.

[–] IncognitoErgoSum@kbin.social 0 points 1 year ago

You need to do your own homework. I'm not doing it for you. What I will do is lay this to rest:

https://en.wikipedia.org/wiki/Stable_Diffusion

Stable Diffusion is a latent diffusion model, a kind of deep generative artificial neural network. Its code and model weights have been released publicly [...]

https://jalammar.github.io/illustrated-stable-diffusion/

The image information creator works completely in the image information space (or latent space). We’ll talk more about what that means later in the post. This property makes it faster than previous diffusion models that worked in pixel space. In technical terms, this component is made up of a UNet neural network and a scheduling algorithm.

[...]

With this we come to see the three main components (each with its own neural network) that make up Stable Diffusion:

  • [...]

https://stable-diffusion-art.com/how-stable-diffusion-work/

The idea of reverse diffusion is undoubtedly clever and elegant. But the million-dollar question is, “How can it be done?”

To reverse the diffusion, we need to know how much noise is added to an image. The answer is teaching a neural network model to predict the noise added. It is called the noise predictor in Stable Diffusion. It is a U-Net model. The training goes as follows.

[...]

It is done using a technique called the variational autoencoder. Yes, that’s precisely what the VAE files are, but I will make it crystal clear later.

The Variational Autoencoder (VAE) neural network has two parts: (1) an encoder and (2) a decoder. The encoder compresses an image to a lower dimensional representation in the latent space. The decoder restores the image from the latent space.

https://www.pcguide.com/apps/how-does-stable-diffusion-work/

Stable Diffusion is a generative model that uses deep learning to create images from text. The model is based on a neural network architecture that can learn to map text descriptions to image features. This means it can create an image matching the input text description.

https://www.vegaitglobal.com/media-center/knowledge-base/what-is-stable-diffusion-and-how-does-it-work

Forward diffusion process is the process where more and more noise is added to the picture. Therefore, the image is taken and the noise is added in t different temporal steps where in the point T, the whole image is just the noise. Backward diffusion is a reversed process when compared to forward diffusion process where the noise from the temporal step t is iteratively removed in temporal step t-1. This process is repeated until the entire noise has been removed from the image using U-Net convolutional neural network which is, besides all of its applications in machine and deep learning, also trained to estimate the amount of noise on the image.

So, I'll have to give you that you're trivially right that Stable Diffusion does use a Markov Chain, but as it turns out, I had the same misconception as you did, that that was some sort of mathematical equation. A markov chain is actually just a process where each step depends only on the step immediately before it, and it most certainly doesn't mean that you're right about Stable Diffusion not using a neural network. Stable Diffusion works by feeding the prompt and partly denoised image into the neural network over some given number of steps (it can do it in a single step, although the results are usually pretty messy). That in and of itself is a Markov chain. However, the piece that's actually doing the real work (that essentially does a Rorschach test over and over) is a neural network.