The point that was being made was that public available data includes a whole lot amount of copyrighted data to begin with and its pretty much impossible to filter it out.
Grand example, the Eiffel tower in Paris is not copyright protected, but the lights on it are so you can only using pictures of the Eiffel tower during the day, if the picture itself isn’t copyright protected by the original photographer. Copyright law has all these complex caveat and exception that make it impossible to tell in glance whether or not it is protected.
This in turn means, if AI cannot legally train on copyrighted materials it finds online without paying huge sums of money then effectively only mega corporation who can pay copyright fines as cost of business will be able to afford training decent AI.
The only other option to produce any ai of such type is a very narrow curated set of known materials with a public use license but that is not going to get you anything competent on its own.
EDIT: In case it isn’t clear i am clarifying what i understood from Grimy@lemmy.world comment, not adding to it.
So then we as a society aren’t ready to untangle the mess of our infancy in the digital age. ChatGPT isn’t something we must have at all costs, it’s something we should have when we can deploy it while still respecting the rights of people who have made the content being used to train it.
I would go even further and say that we should have it until we can be sure it will respect others’ rights. All kind of rights, not only Copyright. Unlike Bing at the beginning, with all it’s bullying and menaces, or Chatgpt regurgitating private information gathered from God knows where.
The problem with waiting is the arms race with other governments. I feel it’s similar to fossil fuels, but all governments need to take the risk of being disadvantaged. Damned prisoner’s dilemma.
I didn’t want any of this shit. IDGAF if we don’t have AI. I’m still not sure the internet actually improved anything, let alone what the benefits of AI are supposed to be.
Machine learning techniques are often thought of as fancy function approximation tools (i.e. for regression and classification problems). They are tools that receive a set of values and spit out some discrete or possibly continuous prediction value.
One use case is that there are a lot of really hard+important problems within CS that we can’t solve efficiently exactly (lookup TSP, SOP, SAT and so on) but that we can solve using heuristics or approximations in reasonable time. Often the accuracy of the heuristic even determines the efficiency of our solution.
Additionally, sometimes we want predictions for other reasons. For example, software that relies on user preference, that predicts home values, that predicts the safety of an engineering plan, that predicts the likelihood that a person has cancer, that predicts the likelihood that an object in a video frame is a human etc.
These tools have legitamite and important use cases it’s just that a lot of the hype now is centered around the dumbest possible uses and a bunch of idiots trying to make money regardless of any associated ethical concerns or consequences.
It’s not like all this data was randomly dumped at the AIs. For data sets to serve as good training materials they need contextual information so that the AI can discern patterns and replicate them when prompted.
We see this when you can literally prompt AIs with whose style you want it to emulate. Meaning that the data it was fed had such information.
Midjourney is facing extra backlash from artists after a spreadsheet was leaked containing a list of artist styles their AI was trained on. Meaning they can keep track of it and they trained the AI with those artists’ works deliberately. They simply pretend this is impossible to figure out so that they might not be liable to seek permission and compensate the artists whose works were used.
I clarified the comment above which was misunderstood, whether it makes a moral/sane argument is subjective and i am not covering that.
I am not sure why you think there is a claim that openAI is trying to make companies pay, on the contrary the comment i was clarifying (so not my opinion/words) states that openAI is making an argument that anyone should be able to use copyrighted materials for free to train AI.
The costs of running an online service like chatgpt is wildly besides the argument presented. You can run your own open source large language models at home about as well as you can run Bethesda’s Starfield on a same spec’d PC
Those Open source large language models are trained on the same collections of data including copyrighted data.
The logic being used here is:
If It becomes globally forbidden to train AI with copyrighted materials or there is a large price or fine in order to use them for training then the Non-Corporate, Free, Open Source Side of AI will perish or have to go underground while to the For-Profit mega corporations will continue exploit and train ai as usual because they can pay to settle in court.
The Ethical dilemma as i understand it is:
Allowing Ai to train for free is a direct threat towards creatives and a win for BigProfit Enthertainment, not allowing it to train to free is treat to public democratic AI and a win for BigTech merging with BigCrime
Allowing Ai to train for free is a direct threat towards creatives
No. Many creatives fear that AI allows anyone to do what they do, lowering the skill premium they can charge. That doesn’t depend on free training.
Some seem to feel that paying for training will delay AI deployment for some years, allowing the good times to continue (until they retire or die?)
But afterward, you have to ask who’s paying for the extra cost when AI is a normal tool for creatives? Where does the money come from to pay the rent to property owners? Obviously the general public will pay a part through higher prices. But I think creatives may bear the brunt, because it’s the tools of their trade that are more expensive and I don’t think all of that cost can be passed on.
I don’t think lowering the skill level is something we will need to worry about as over time this actually trickles up, A Creative professional trained with AI tools will almost always top a Amateur using the same tools.
The real issue is Style. If you are an Artist with a very recognizable specific style, and you make your money trough commissions you are basically screwed.
Many Artists feature a personal style and while borrowing peoples style is common (disney-esque) it’s usually not a problem because within a unique and diverse human mind it rarely results in unintentional latent copying.
I think, in the short run, some have reason to worry about their skills. AI does make digital skills more important and manual drawing skills less so.
OTOH, I don’t think it’s reasonable to worry about styles. Go to aliexpress or some such place and look for paintings. They offer cheap “handmade” paintings and replicas of famos works. They don’t offer novel paintings in someone else’s style. I don’t believe there is any demand for that.
The people i worry about most are the independent “starvin” artists you may find at conventions or working from their social media profile.
They often do personalized commissioned pieces, what they are essential selling is the clients idea in a chosen style which they have mastered.
That entire biz is at risk of going away. I believe the talent of those artists is still valuable though so it is my hope they can go on and make higher level art using the tools they are fighting now.
That is very well put, I really wish I could have started with that.
Though I envision it as a loss for BigProfit Enthertainment since I see this as a real boon for the indie gaming, animation and eventually filmmaking industry.
You can run your own open source large language models at home about as well as you can run Bethesda’s Starfield on a same spec’d PC
…
Yes, you can download an executable of a chatbot lol.
That’s different than running something remotely like even OpenAI.
The more it has to reference, the more the system scales up. Not just storage, but everything else.
Like, in your example of video games it would be more like stripping down a PS5 game of all the assets, then playing it on a NES at 1 frame per five minutes.
You’re not only wildly overestimating chatbots ability, you’re doing that while drastically underestimating the resources needed.
Edit:
I think you literally don’t know what people are talking about…
Do you think people are talking about AI image generators?
I am talking about generative AI, be it text or image both have a challenge with copyrighted material.
“executable of a chatbot”
lol, aint you cute
“example of video games”
Are you refering to my joke?
I am far from overestimating capacity, Starfield runs mediocre on a modern gaming system compared to other games.
The Vicuna 13b llm runs mediocre on the same system compared with gpt 3.5. To this date there is no local model that i would trust for professional use and chatgpt 3.5 doesnt hit that level either.
But it remains a very interesting, rapidly evolving technology that i hope receives as much future open source support as possible.
“I think you literally don’t know what people are talking about”
I hate to break it to you but you’re embarrassing yourself.
I presume you must believe the the following lemmy community and resources to be typed up by a group of children, either that or your just naive.
HOT TAKE: Hugging face is run by people who are really into chatbots but dont understand it in the slightest.
I have been patient and friendly so far but your tone has been nothing but dismissive.
you cannot have a nuanced conversation about AI while excluding the entire Open Source field within it. That’s simply unreasonable and i plore you to ask others because i know you wont take my word for it.
I’m not sure if someone else has brought this up, but I could see OpenAI and other early adopters pushing for tighter controls of training data as a means to be the only players in town. You can’t build your own competing AI because you won’t have the same amount of data as us and we’ll corner the market.
Eh, the issue is lots of people wouldn’t be willing to sell tho.
Like, you think an author wants the chatbot to read their collected works and use that? Regardless of if it’s quoting full texts or “creating” text in their style.
No author is going to want that.
And if it’s up to publishers, they likely won’t either. Why take one small payday if that could potentially lead to loss of sales a few years down the row.
It’s not like the people making the chatbits just need to buy a retail copy of the text to be in the legal clear.
It’s current and it’s the only open source project that’s under direct threat? I am both a fan of open source and of generative AI, not sure what that changes in the validity of my arguments.
This isn’t a gotcha but pure rhetoric, which is on par with you. Attack my arguments, or just ignore me the moment it becomes clear you can’t insult yourself out of a debate like you did last time.
I’m not even sure what exactly you are implying but I am not impressed.
That is literally rhetoric. I could say the same about you and never mentioning artists except when it’s related to AI. But I don’t, I pick your weak arguments apart like an adult instead.
And using publicly available data to train gets you a shitty chatbot…
Hell, even using copyrighted data to train isn’t that great.
Like, what do you even think they’re doing here for your conspiracy?
You think OpenAI is saying they should pay for the data? They’re trying to use it for free.
Was this a meta joke and you had a chatbot write your comment?
if someone said this to me I’d cry
The point that was being made was that public available data includes a whole lot amount of copyrighted data to begin with and its pretty much impossible to filter it out. Grand example, the Eiffel tower in Paris is not copyright protected, but the lights on it are so you can only using pictures of the Eiffel tower during the day, if the picture itself isn’t copyright protected by the original photographer. Copyright law has all these complex caveat and exception that make it impossible to tell in glance whether or not it is protected.
This in turn means, if AI cannot legally train on copyrighted materials it finds online without paying huge sums of money then effectively only mega corporation who can pay copyright fines as cost of business will be able to afford training decent AI.
The only other option to produce any ai of such type is a very narrow curated set of known materials with a public use license but that is not going to get you anything competent on its own.
EDIT: In case it isn’t clear i am clarifying what i understood from Grimy@lemmy.world comment, not adding to it.
So then we as a society aren’t ready to untangle the mess of our infancy in the digital age. ChatGPT isn’t something we must have at all costs, it’s something we should have when we can deploy it while still respecting the rights of people who have made the content being used to train it.
I would go even further and say that we should have it until we can be sure it will respect others’ rights. All kind of rights, not only Copyright. Unlike Bing at the beginning, with all it’s bullying and menaces, or Chatgpt regurgitating private information gathered from God knows where.
The problem with waiting is the arms race with other governments. I feel it’s similar to fossil fuels, but all governments need to take the risk of being disadvantaged. Damned prisoner’s dilemma.
I didn’t want any of this shit. IDGAF if we don’t have AI. I’m still not sure the internet actually improved anything, let alone what the benefits of AI are supposed to be.
Machine learning techniques are often thought of as fancy function approximation tools (i.e. for regression and classification problems). They are tools that receive a set of values and spit out some discrete or possibly continuous prediction value.
One use case is that there are a lot of really hard+important problems within CS that we can’t solve efficiently exactly (lookup TSP, SOP, SAT and so on) but that we can solve using heuristics or approximations in reasonable time. Often the accuracy of the heuristic even determines the efficiency of our solution.
Additionally, sometimes we want predictions for other reasons. For example, software that relies on user preference, that predicts home values, that predicts the safety of an engineering plan, that predicts the likelihood that a person has cancer, that predicts the likelihood that an object in a video frame is a human etc.
These tools have legitamite and important use cases it’s just that a lot of the hype now is centered around the dumbest possible uses and a bunch of idiots trying to make money regardless of any associated ethical concerns or consequences.
It doesn’t matter what you want. What matters is if corporations can extract $ from you, gain an efficiency, or cut their workforce using it.
That’s what the drive for AI is all about.
No doubt.
A perfectly valid stance to take.
You don’t have to use it. You can even disconnect from the internet completely.
Whats the benefit of stopping me from using it?
It’s not like all this data was randomly dumped at the AIs. For data sets to serve as good training materials they need contextual information so that the AI can discern patterns and replicate them when prompted.
We see this when you can literally prompt AIs with whose style you want it to emulate. Meaning that the data it was fed had such information.
Midjourney is facing extra backlash from artists after a spreadsheet was leaked containing a list of artist styles their AI was trained on. Meaning they can keep track of it and they trained the AI with those artists’ works deliberately. They simply pretend this is impossible to figure out so that they might not be liable to seek permission and compensate the artists whose works were used.
That’s insane logic…
Like you’re essentially saying I can copy/paste any article without a paywall to my own blog and sell adspace on it…
And your still saying OpenAI is trying to make AI companies pay?
Like, do you think AI runs off free cloud services? The hardware is insanely expensive.
And OpenAI is trying to argue the opposite, that AI companies shouldn’t have to pay to use copyrighted works.
You have zero idea what is going on, but you are really confident you do
I clarified the comment above which was misunderstood, whether it makes a moral/sane argument is subjective and i am not covering that.
I am not sure why you think there is a claim that openAI is trying to make companies pay, on the contrary the comment i was clarifying (so not my opinion/words) states that openAI is making an argument that anyone should be able to use copyrighted materials for free to train AI.
The costs of running an online service like chatgpt is wildly besides the argument presented. You can run your own open source large language models at home about as well as you can run Bethesda’s Starfield on a same spec’d PC
Those Open source large language models are trained on the same collections of data including copyrighted data.
The logic being used here is:
The Ethical dilemma as i understand it is:
No. Many creatives fear that AI allows anyone to do what they do, lowering the skill premium they can charge. That doesn’t depend on free training.
Some seem to feel that paying for training will delay AI deployment for some years, allowing the good times to continue (until they retire or die?)
But afterward, you have to ask who’s paying for the extra cost when AI is a normal tool for creatives? Where does the money come from to pay the rent to property owners? Obviously the general public will pay a part through higher prices. But I think creatives may bear the brunt, because it’s the tools of their trade that are more expensive and I don’t think all of that cost can be passed on.
I don’t think lowering the skill level is something we will need to worry about as over time this actually trickles up, A Creative professional trained with AI tools will almost always top a Amateur using the same tools.
The real issue is Style. If you are an Artist with a very recognizable specific style, and you make your money trough commissions you are basically screwed. Many Artists feature a personal style and while borrowing peoples style is common (disney-esque) it’s usually not a problem because within a unique and diverse human mind it rarely results in unintentional latent copying.
I think, in the short run, some have reason to worry about their skills. AI does make digital skills more important and manual drawing skills less so.
OTOH, I don’t think it’s reasonable to worry about styles. Go to aliexpress or some such place and look for paintings. They offer cheap “handmade” paintings and replicas of famos works. They don’t offer novel paintings in someone else’s style. I don’t believe there is any demand for that.
The people i worry about most are the independent “starvin” artists you may find at conventions or working from their social media profile. They often do personalized commissioned pieces, what they are essential selling is the clients idea in a chosen style which they have mastered. That entire biz is at risk of going away. I believe the talent of those artists is still valuable though so it is my hope they can go on and make higher level art using the tools they are fighting now.
That is very well put, I really wish I could have started with that.
Though I envision it as a loss for BigProfit Enthertainment since I see this as a real boon for the indie gaming, animation and eventually filmmaking industry.
It’s definitely overall quite a messy situation.
…
Yes, you can download an executable of a chatbot lol.
That’s different than running something remotely like even OpenAI.
The more it has to reference, the more the system scales up. Not just storage, but everything else.
Like, in your example of video games it would be more like stripping down a PS5 game of all the assets, then playing it on a NES at 1 frame per five minutes.
You’re not only wildly overestimating chatbots ability, you’re doing that while drastically underestimating the resources needed.
Edit:
I think you literally don’t know what people are talking about…
Do you think people are talking about AI image generators?
No one else is…
I think you’re confusing training it with running it. After it’s trained, you can run it on much weaker hardware.
The issue is it reproducing copyrighted works verbatim…
It can’t do that unless it contains the entire text to begin with…
I am talking about generative AI, be it text or image both have a challenge with copyrighted material.
Are you refering to my joke?
I am far from overestimating capacity, Starfield runs mediocre on a modern gaming system compared to other games. The Vicuna 13b llm runs mediocre on the same system compared with gpt 3.5. To this date there is no local model that i would trust for professional use and chatgpt 3.5 doesnt hit that level either.
But it remains a very interesting, rapidly evolving technology that i hope receives as much future open source support as possible.
I presume you must believe the the following lemmy community and resources to be typed up by a group of children, either that or your just naive.
https://lemmy.world/c/fosai
https://www.fosai.xyz/
https://github.com/huggingface/transformers
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
https://huggingface.co/microsoft/phi-2 & https://www.microsoft.com/en-us/research/blog/phi-2-the-surprising-power-of-small-language-models/
https://www.theguardian.com/technology/2023/may/05/google-engineer-open-source-technology-ai-openai-chatgpt
Or…
I could just block some of the people who are really really into chatbots, but don’t understand it in the slightest.
I think that might be more productive than reading a bunch of stuff from other people who don’t understand it.
HOT TAKE: Hugging face is run by people who are really into chatbots but dont understand it in the slightest.
I have been patient and friendly so far but your tone has been nothing but dismissive.
you cannot have a nuanced conversation about AI while excluding the entire Open Source field within it. That’s simply unreasonable and i plore you to ask others because i know you wont take my word for it.
Farewell
Hey man, that’s damn hurtful
I’m not sure if someone else has brought this up, but I could see OpenAI and other early adopters pushing for tighter controls of training data as a means to be the only players in town. You can’t build your own competing AI because you won’t have the same amount of data as us and we’ll corner the market.
If the data has to be paid for, openAI will gladly do it with a smile on their face. It guarantees them a monopoly and ownership of the economy.
Paying more but having no competition except google is a good deal for them.
Eh, the issue is lots of people wouldn’t be willing to sell tho.
Like, you think an author wants the chatbot to read their collected works and use that? Regardless of if it’s quoting full texts or “creating” text in their style.
No author is going to want that.
And if it’s up to publishers, they likely won’t either. Why take one small payday if that could potentially lead to loss of sales a few years down the row.
It’s not like the people making the chatbits just need to buy a retail copy of the text to be in the legal clear.
The publisher’s will absolutely sell imo. They just publish, the book will be worth the same with or without the help of AI to write it.
I guess there is a possibility that people start replacing bought books with personalized book llm outputs but that strikes me as unlikely.
deleted
deleted
It’s current and it’s the only open source project that’s under direct threat? I am both a fan of open source and of generative AI, not sure what that changes in the validity of my arguments.
This isn’t a gotcha but pure rhetoric, which is on par with you. Attack my arguments, or just ignore me the moment it becomes clear you can’t insult yourself out of a debate like you did last time.
I’m not even sure what exactly you are implying but I am not impressed.
deleted
That is literally rhetoric. I could say the same about you and never mentioning artists except when it’s related to AI. But I don’t, I pick your weak arguments apart like an adult instead.
deleted