this post was submitted on 02 Jan 2025
18 points (72.5% liked)
Asklemmy
44260 readers
1508 users here now
A loosely moderated place to ask open-ended questions
Search asklemmy ๐
If your post meets the following criteria, it's welcome here!
- Open-ended question
- Not offensive: at this point, we do not have the bandwidth to moderate overtly political discussions. Assume best intent and be excellent to each other.
- Not regarding using or support for Lemmy: context, see the list of support communities and tools for finding communities below
- Not ad nauseam inducing: please make sure it is a question that would be new to most members
- An actual topic of discussion
Looking for support?
Looking for a community?
- Lemmyverse: community search
- sub.rehab: maps old subreddits to fediverse options, marks official as such
- !lemmy411@lemmy.ca: a community for finding communities
~Icon~ ~by~ ~@Double_A@discuss.tchncs.de~
founded 5 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Exactly, an equitable solution could be to pay royalties to artists that had their work stolen to train these algorithms. That, however, would require any of the generative algorithms to be operating at a profit, which they absolutely are not.
And it would require the LLM owners to admit to stealing that content.
That too
One more thing: if you want to use public data, your AI needs to be open source (not just the software around it, the actual models that do the AI stuff needs to be available for anyone to run on their own system) and all the works generated with it public domain. The public owns your AI at that point. Personally, if you don't want to pay me, then let me have a stake in the AI my data helped create.
That's a good point.
Copying is not theft
In this case it absolutely is.
Google it