"because it's supposedly "impossible" for the company to train its artificial intelligence models — and continue growing its multi-billion-dollar-business — without them."
O no! Poor richs cant get more rich fast enough :(
This is a most excellent place for technology news and articles.
"because it's supposedly "impossible" for the company to train its artificial intelligence models — and continue growing its multi-billion-dollar-business — without them."
O no! Poor richs cant get more rich fast enough :(
Copyright is a pain in the ass, but Sam Altman is a bigger pain in the ass. Send him to prison and let him rot. Then put his tears in a cup and I'll drink them
Oh no. Anyway...
Shut it down then and stop stealing other peoples shit
that guy in that picture looks like the "unwanted house guest" from those memes from 10 years ago
Wow, that's a shame. Anyway, take all his money and throw him in a ditch someplace.
What irks me most about this claim from OpenAI and others in the AI industry is that it's not based on any real evidence. Nobody has tested the counterfactual approach he claims wouldn't work, yet the experiments that came closest--the first StarCoder LLM and the CommonCanvas text-to-image model--suggest that, in fact, it would have been possible to produce something very nearly as useful, and in some ways better, with a more restrained training data curation approach than scraping outbound Reddit links.
All that aside, copyright clearly isn't the right framework for understanding why what OpenAI does bothers people so much. It's really about "data dignity", which is a relatively new moral principle not yet protected by any single law. Most people feel that they should have control over what data is gathered about their activities online, as well as what is done with those data after it's been collected, and even if they publish or post something under a Creative Commons license that permits derived uses of their work, they'll still get upset if it's used as an input to machine learning. This is true even if the generative models thereby created are not created for commercial reasons, but only for personal or educational purposes that clearly constitute fair use. I'm not saying that OpenAI's use of copyrighted work is fair, I'm just saying that even in cases where the use is clearly fair, there's still a perceived moral injury, so I don't think it's wise to lean too heavily on copyright law if we want to find a path forward that feels just.
It's impossible for me to make money without robbing a bank, please let me do that parliament it would be so funny
Honestly, copyright is shit. It is created on the basis of an old way of doing things. That is, where big editors and big studios make mass productions of physical copies of a said 'product'. George R. R. Martin , Warner Studios & co are rich. Maybe they have everything to lose without their copy'right' but that isn't the population's problem. We live in an era where everything is digital and easily copiable and we might as well start acting like it.
I don't care if Sam Altman is evil, this discussion is fundamental.
How did GRRM get rich again?
oh yeah he sold books he worked on for decades, totally the same WB.
What kind of a pathetic statement is that ?
We can't make money paying for "AI", going to theaters, or paying for streaming services.
So I guess everybody gets a piracy!
Sounds like they need better bootstraps.
Or at least a business model.
Idk, usually people shut down their business if it can't make a profit...
Aww poor shit company and their poor money problems.
“Too fucking bad”
So I got a crazy idea - hear me out - how about we just abolish copyright completely, for everyone?
I mean, it works in China pretty well.
https://en.wikipedia.org/wiki/Intellectual_property_in_China
Looks like there are still copyright laws in China. What are you on about?
The internet has been primarily derivative content for a long time. As much as some haven't wanted to admit it. It's true. These fancy algorithms now take it to the exponential factor.
Original content had already become sparsely seen anymore as monetization ramped up. And then this generation of AI algorithms arrived.
The several years before prior to LLMs becoming a thing, the internet was basically just regurgitating data from API calls or scraping someone else's content and representing it in your own way.
Ok... Is that supposed to be a good reason?