I'm fine with this. "We can't succeed without breaking the law" isn't much of an argument.
Do I think the current copyright laws around the world are fine? No, far from it.
But why do they merit an exception to the rules that will make them billions, but the rest of us can be prosecuted in severe and dramatic fashion for much less. Try letting the RIAA know you have a song you've downloaded on your PC that you didn't pay for - tell them it's for "research and training purposes", just like AI uses stuff it didn't pay for - and see what I mean by severe and dramatic.
It should not be one rule for the rich guys to get even richer and the rest of us can eat dirt.
Figure out how to fix the laws in a way that they're fair for everyone, including figuring out a way to compensate the people whose IP you've been stealing.
Until then, deal with the same legal landscape as everyone else. Boo hoo
"We can't succeed without breaking the law. We can't succeed without operating unethically."
I'm so sick of this bullshit. They pretend to love a free market until it's not in their favor and then they ask us to bend over backwards for them.
Too many people think they're superior. Which is ironic, because they're also the ones asking for handouts and rule bending. If you were superior, you wouldn't need all the unethical things that you're asking for.
Sam Altman is a grifter, but on this topic he is right.
The reality is, that IP laws in their current form hamper innovation and technological development. Stephan Kinsella has written on this topic for the past 25 years or so and has argued to reform the system.
Here in the Netherlands, we know that it's true. Philips became a great company because they could produce lightbulbs here, which were patented in the UK. We also had a booming margarine business, because we weren't respecting British and French patents and that business laid the foundation for what became Unilever.
And now China is using those exact same tactics to build up their industry. And it gives them a huge competitive advantage.
A good reform would be to revert back to the way copyright and patent law were originally developed, with much shorter terms and requiring a significant fee for a one time extension.
The current terms, lobbied by Disney, are way too restrictive.
If giant megacorporations can benefit by ignoring copyright, us mortals should be able to as well.
Until then, you have the public domain to train on. If you don't want AI to talk like the 1920s, you shouldn't have extended copyright and robbed society of a robust public domain.
That's like calling stealing from shops essential for my existence and it would be "over" for me if they stop me. The shit these clowns say is just astounding. It's like they have no morals and no self awareness and awareness for people around them.
The only way this would be ok is if openai was actually open. make the entire damn thing free and open source, and most of the complaints will go away.
Copyrights should have never been extended longer than 5 years in the first place, either remove draconian copyright laws or outlaw LLM style models using copyrighted material, corpos can't have both.
If I had to pay tuition for education (buying text books, pay for classes and stuff), then you have to pay me to train your stupid AI using my materials.
TLDR: "we should be able to steal other people's work, or we'll go crying to daddy Trump. But DeepSeek shouldn't be able to steal from the stuff we stole, because China and open source"
LLM algorithms can be maintained and sold to corpos to scrape their own data so they can use them for in house tools, or re-sell them to their own clients.
Open Source LLMs can be made available for end users to do the same with their own data, or scrape whats available in the public domain for whatever they want so long as they don't re-sell
If your business model is predicated on breaking the law then you don't deserve to exist.
You can't send people to prison for 5 years and charge them $100,000 for downloading a movie and then turn around and let big business do it for free because they need to "train their AI model" and call one of thief but not the other...
Look we may have driven Aaron Swartz to suicide for doing basically the same thing on a smaller scale, but dammit we are getting very rich of this. And, if we are getting rich, then it is okay to break the law while actively fucking over actually creative people. Trust us. We are tech bros and we know what is best for you is for us to become incredibly rich and out of touch. You need us.
At the end of the day the fact that openai lost their collective shit when a Chinese company used their data and model to make their own more efficient model is all the proof I need they don't care about being fair or equitable when they get mad at people doing the exact thing they did and would aggressively oppose others using their own work to advance their own.
Why training openai with literally millions of copyrighted works is fair use, but me downloading an episode of a series not available in any platform means years of prison?
Sounds like another way of saying "there actually isn't a profitable business in this."
But since we live in crazy world, once he gets his exemption to copyright laws for AI, someone needs to come up with a good self hosted AI toolset that makes it legal for the average person to pirate stuff at scale as well.
This is basically a veiled admission that OpenAI are falling behind in the very arms race they started. Good, fuck Altman. We need less ultra-corpo tech bro bullshit in prevailing technology.
God forbid you offer to PAY for access to works that people create like everyone else has to. University students have to pay out the nose for their books that they "train" on, why can't billion dollar AI companies?
I'm fine for them to use copyrighted material, provided that everyone can do the same without reprecautions
Fuck double standards. Fuck IP. People should have access to knowledge without having to pay.
What if we had taken the billions of dollars invested in AI and invested that into public education instead?
Imagine the return on investment of the information being used to train actual humans who can reason and don’t lie 60% of the time instead of using it to train a computer that is useless more than it is useful.
I have conflicting feelings about this whole thing. If you are selling the result of training like OpenAI does (and every other company), then I feel like it’s absolutely and clearly not fair use. It’s just theft with extra steps.
On the other hand, what about open source projects and individuals who aren’t selling or competing with the owners of the training material? I feel like that would be fair use.
What keeps me up at night is if training is never fair use, then the natural result is that AI becomes monopolized by big companies with deep pockets who can pay for an infinite amount of random content licensing, and then we are all forever at their mercy for this entire branch of technology.
The practical, socioeconomic, and ethical considerations are really complex, but all I ever see discussed are these hard-line binary stances that would only have awful corporate-empowering consequences, either because they can steal content freely or because they are the only ones that will have the resources to control the technology.
If AI gets to use copyrighted material for free and makes a profit off of the results, that means piracy is 1000% Legal.
Excuse me while I go and download a car!!
How many pages has a human author read and written before they can produce something worth publishing? I’m pretty sure that’s not even a million pages. Why does an AI require a gazillion pages to learn, but the quality is still unimpressive? I think there’s something fundamentally wrong with the way we teach these models.
I hope generative AI obliterates copyright. I hope that its destruction is so thorough that we either forget it ever existed or we talk about it in disgust as something that only existed in stupider times.
If I'm using "AI" to generate subtitles for the "community" is ok if i have a large "datastore" of "licensable media" stored locally to work off of right?
I need a seamstress AI to take over 10 million seamstress robots so I don't have to pay 100million seamstresses for fruit of the loom underwear.... Could you tech it how to do double well and then back up at each end with some zigzags? For free? I mean everyone knows zigzag!
This is exactly what social media companies have been doing for a while (it’s free, yes) they use your data to train their algorithms to squeeze more money out of people. They get a tangible and monetary benefit from our collective data. These AI companies want to train their AI on our hard work and then get monetary benefit off of it. How is this not seen as theft or even if they are not doing it just yet…how is it not seen as an attempt at theft?
How come people (not the tech savvy) are unable to see how they are being exploited? These companies are not currently working towards any UBI bills or policies in governments that I am aware of. Since they want to take our work, and use it to get rich and their investors rich why do they think they are justified in using people’s work? It just seems so slime-y.
Sad to see you leave (not really, tho'), love to watch you go!
Edit: I bet if any AI developing company would stop acting and being so damned shady and would just ASK FOR PERMISSION, they'd receive a huge amount of data from all over. There are a lot of people who would like to see AGI become a real thing, but not if it's being developed by greedy and unscrupulous shitheads. As it stands now, I think the only ones who are actually doing it for the R&D and not as eye-candy to glitz away people's money for aesthetically believable nonsense are a handful of start-up-likes with (not in a condescending way) kids who've yet to have their dreams and idealism trampled.
What's wrong with the sentiment expressed in the headline? AI training is not and should not be considered fair use. Also, copyright laws are broken in the west, more so in the east.
We need a global reform of copyright. Where copyrights can (and must) be shared among all creators credited on a work. The copyright must be held by actual people, not corporations (or any other collective entity), and the copyright ends after 30 years or when the all rights holders die, whichever happens first. That copyright should start at the date of initial publication. The copyright should be nontransferable but it should be able to be licensed to any other entity only with a majority consent of all rights holders. At the expiration of the copyright the work in question should immediately enter the public domain.
And fair use should be treated similarly to how it is in the west, where it's decided on a case-by-case basis, but context and profit motive matter.
Musk has an AI project. Techbros have deliberately been sucking up to Trump. I’m pretty sure AI training will be declared fair use and copyright laws will remain the same for everybody else.
Its simple really. We need to steal from humans who make things to train our computers to make things so the computers can replace the humans who make things and we dont want to pay the humans who made the original things because they will be replaced soon enough anyway. Easy peasy. Do you guys even capitalism?
Good. If I ever published anything, I would absolutely not want it to be pirated by AI so some asshole can plagiarize it later down the line and not even cite their sources.
Guarantee their plan is to blow through copyright laws to create a monopoly fiefdom, close the door behind them and demand that copyright is used to protect the work their LLM creates.
I don’t think they’re wrong in saying that if they aren’t allowed to train on copyrighted works then they will fall behind. Maybe I missed it in the article, but Japan for example has that exact law (use of copyright to train generative AI is allowed).
Personally I think we need to give them somewhat of an out by letting them do it but then taxing the fuck out of the resulting product. “You can use copyrighted works for training but then 50% of your profits are taxed”. Basically a recognition that the sum of all copyrighted works is a societal good and not just an individual copyright holders.
If I don't want people copying it, people shouldn't be copying it. I don't care if it's been 500 years. It's my book.
This is a weird thread. Lots of people for artists losing control of their creations quickly while simultaneously against artist creations being used by others without consent. Just my perspective but why should artists lose control of their own creations at all? The problem in copyright is tech companies doing patent thickets; not artists.
Even artistic creations held by corporations. Waiting for Marvel stuff to hit public domain to publish a bunch of Marvel novels since they can't protect their creations any more? Why is that acceptable? If someone creates something and doesn't want it stolen, I don't give a fuck what the law says, stealing it is theft. The thief should instead be using Marvel stuff as inspiration as they make their own universe; not just waiting an amount of time before stealing someone else's creation without consent. It isn't holding progress back at all to make novel artistic creations instead of steal others. Art = very different from tech.
when I publish a book, to steal it is consenting to be Luigi'd; no matter how long ago it came out.
It's so wild how laws just have no idea what to do with you if you just add one layer of proxy.
"Nooo I'm not stealing and plagerizing, it's the AI doing it!"
Ip should solely be with the creator and not the corporation that owns that creator. A lot of problems in stems is IP held hostage by the corporations and by publishing companies of research papers
OpenAI also had no clue on recreation the happy little accident that gave them chatGPT3. That's mostly because their whole thing was using a simple model and brute forcing it with more data, more power, more nodes and then even more data and power until it produced results.
As expected, this isn't sustainable. It's beyond the point of decreasing returns. But Sam here has no idea on how to fix that with much better models so goes back to the one thing he knows: more data needed, just one more terabyte bro, ignore the copyright!
And now he's blaming the Chinese into forcing him to use even more data.
I mean if they pay for it like everyone else does I don't think it is a problem. Yes it will cost you billions and billions to do it correctly, but then you basically have the smartest creature on earth (that we know of) and you can replicate/improve on it in perpetuity. We still will have to pay you licensing fees to use it in our daily lives, so you will be making those billions back.
Now I would say let them use anything that is old and freeware, textbooks, etc. government owned stuff - we sponsored it with our learning, taxes - so we get a percentage in all AI companies. Humanity gets a 51% stake in any AI business using humanity's knowledge, so we are then free to vote on how the tech is being used and we have a controlling share, also whatever price is set, we get half of it back in taxes at the end of the year. The more you use it the more you pay and the more you get back.
In the early 80s I used to have fantasies about having a foster robot android that my family was teaching how to be a person. Oh the amusing mix-ups we got into! We could just do that. Train on experiential reality instead of on the dim cultural reflection of reality.
As far as the ai industry has already broken copyright laws. It will not be actually intelligent for a long time. Just like crypto this seems like a global scam that has squandered resources for a dream of a free workforce. Instead of working together to try and create an ai there are lots of technology companies doing the same ineffective bull 🤔
Perhaps this is just a problem with the way the model works. Always requiring new data and unable to use current data, to ponder and expand upon while making new connections about ideas that influenced the author… LLM’s are a smoke and mirrors show, not a real intelligence.
chatgpt is stagnant, the newest model was lackluster despite using way more resources and costing a shitload of cash, Altman is floundering and on his way out he’s going to try do some lobbying bullshit
Copyright is bullshit and honestly if it disappeared it would help small creators more than anything but openai is not a small creator and guaranteed they will lobby for only huge corps like them to get such an exception. You and I will still get sued to shit by disney or whoever for daring to make $500 off of a shitty project that used some sample or something while meta and openai get free reign to steal the entirety of humanity’s creative output with no recompense