Comment by slidehero
>There's no reason why training on a billion images is any different
You gloss over this as if it's a given. I don't agree. I think you're doing a different thing when you're sampling billions of things equallly.
>There's no reason why training on a billion images is any different
You gloss over this as if it's a given. I don't agree. I think you're doing a different thing when you're sampling billions of things equallly.
> the model reproduces Indiana Jones
the model isn't the one infringing. It's the end user inputting the prompt.
The model itself is not a derivative work, in the same way that an artist and photoshop aren't a derivative work when they reproduce indiana jones's likeness.
That does not seem obvious at all. Fan art and referencing is a thing, and there are plenty of examples of AI creating characters that do not exist anywhere in the training dataset.
That's why I said it's an argument by induction. Where's the limit for it to be different? 10 images? 100? 10000? Where does it stop being copyright infringement and why? Many people have paid heavy fines for much less. I don't think that "a billion images is so unfathomable compared to just one million that it truly is a difference in kind" is a valid response
The root problem is that the model reproduces Indiana Jones instead of creating a new character. This contradicts the statement that the model "learns" and "creates" like a human artist and not merely copies; obviously a human artist would not plagiarize when asked to draw a character.