

Did you respond to the wrong person? The article nor my comment was about one specific AI model.
Did you respond to the wrong person? The article nor my comment was about one specific AI model.
Most likely because if he would, it would remove one of the last barriers Twitter has to being sanctioned even harder by eg. the EU.
This is the inevitable end game of some groups of people trying to make AI usage taboo using anger and intimidation without room for reasonable disagreement. The ones devoid of morals and ethics will use it to their hearts content and would never interact with your objections anyways, and when the general public is ignorant of what it is and what it can really do, people get taken advantage off.
Support open source and ethical usage of AI, where artists, creatives, and those with good intentions are not caught in your legitimate grievances with corporate greed, totalitarians, and the like. We can’t reasonably make it go away, but we can reduce harmful use of it.
While there are spaces that are luckily still looking at it neutrally and objectively, there are definitely leftist spaces where AI hatred has snuck in, even to a reality-denying degree where lies about what AI is or isn’t has taken hold, and where providing facts to refute such things are rejected and met with hate and shunning purely because it goes against the norm.
And I can’t help but agree that they are being played so that the only AI technology that will eventually be feasible will not be open source, and in control of the very companies left learning folks have dislike or hatred for.
The absolute irony
I never claimed anything besides that breakthroughs did happen since you claimed, which is objectively true. You claimed very concretely that AI was the same for over a decade, aka it was the same in at least 2015 if I’m being charitable, all of these things were researched in the last 7-8 years and only became the products as we know them in the last 5 years. (Aka 2020)
Breakthroughs are not a myth. They still happen even when the process is iterative. That page even explains it. The advent of the GAN (2014-2018), which got overtaken by transformers in around 2017 for which GPTs and Diffusion models later got developed on. More hardware is what allowed those technologies to work better and bigger but without those breakthroughs you still wouldnt have the AI boom of today.
ML technology has existed for a while, but it’s wild to claim that the technology pre-2020 is the same. A breakthrough happened.
Oh I agree money talks in the US justice system, but as the page shows, these laws also exist elsewhere, such as in the EU. And even if I or you don’t agree with them, they are still the case law that determines the legality of these things. For me that aligns with my ethical stance as well, but probably not yours.
I never claimed that in this case. As I said in my response: There have been won lawsuits that machines are allowed to index and analyze copyrighted material without infringing on such rights, so long as they only extract objective information, such as what AI typically extracts. I’m not a lawyer, and your jurisdiction may differ, but this page has a good overview: https://blog.apify.com/is-web-scraping-legal/
EDIT: For the US description on that page, it mentions the US case that I referred to: Author’s Guild v Google
Outside of the marketing labels of “artificial intelligence” and “machine learning”, it’s nothing like real intelligence or learning at all.
Generative AI uses artificial neural networks, which are based on how we understand brains to connect information (Biological neural networks). You’re right that they have no self generated input like humans do, but their sense of making connections between information is very similar to that of humans. It doesn’t really matter that they don’t have their own experiences, because they are not trying to be humans, they are trying to be as flexible of a ‘mind’ as possible.
Are you an artist or a creative person?
I see anti-AI people say this stuff all the time too. Because it’s a convenient excuse to disregard an opposing opinion as ‘doesn’t know art’, failing to realize or respect that most people have some kind of creative spark and outlet. And I know it wasn’t aimed at me, but before you think I’m dodging the question, I’m a creative working professionally with artists and designers.
Professional creative people and artists use AI too. A lot. Probably more than laypeople, because to use it well and combine it with other interesting ideas, requires a creative and inventive mind. There’s a reason AI is making it’s way all over media, into movies, into games, into books. And I don’t mean as AI slop, but well-implemented, guided AI usage.
I could ask you as well if you’ve ever studied programming, or studied psychology, as those things would all make you more able to understand the similarities between artificial neural networks and biological neural networks. But I don’t need a box to disregard you, the substance of your argument fails to convince me.
At the end of the day, it does matter that humans have their own experiences to mix in. But AI can also store much, much more influences than a human brain can. That effectively means for everything it makes, there is less of a specific source in there from specific artists.
For example, the potential market effects of generating an automated system which uses people’s artwork to directly compete against them.
Fair use considerations do not apply to works that are so substantially different from any influence, only when copyrighted material is directly re-used. If you read Harry Potter and write your own novel about wizards, you do not have to credit nor pay royalties to JK Rowling, so long as it isn’t substantially similar. Without any additional laws prohibiting such, AI is no different. To sue someone over fair use, you typically do have to prove that it infringes on your work, and so far there have not been any successful cases with that argument.
Most negative externalities from AI come from capitalism: Greedy bosses thinking they can replace true human talent with a machine, plagiarists that use it as a convenient tool to harass specific artists, scammers that use it to scam people. But around that exists an entire ecosystem of people just using it for what it should be used for: More and more creativity.
You picked the wrong thread for a nuanced question on a controversial topic.
But it seems the UK indeed has laws for this already if the article is to believed, as they don’t currently allow AI companies to train on copyrighted material (As per the article). As far as I know, in some other jurisdictions, a normal person would absolutely be allowed to pull a bunch of publicly available information, learn from it, and decide to make something new based on objective information that can be found within. And generally, that’s the rationale AI companies used as well, seeing as there have been landmark cases ruled in the past to not be copyright infringement with wide acceptance for computers analyzing copyrighted information, such as against Google, for indexing copyrighted material in their search results. But perhaps an adjacent ruling was never accepted in the UK (which does seem strange, as Google does operate there). But laws are messy, and perhaps there is an exception somewhere, and I’m certainly not an expert on UK law.
But people sadly don’t really come into this thread to discuss the actual details, they just see a headline that invokes a feeling of “AI Bad”, and so you coming in here with a reasonable question makes you a target. I wholly expect to be downvoted as well.
Yeah I agree that’s a fair enough assumption 😄
Never assumed you did :), but yes, as little assumptions is the best. But as you can already tell, it’s hard to communicate when you take no assumptions when people make explicit statements crafted to dispel assumptions, that are entirely plausible for a hypothetical real person to have.
In fact, your original statement of “They have no doubts. Never occurred to them it might be a joke…”, is in itself a pretty big assumption. Unless, of course. I assume that statement to be a hyperbole, or even satire. But if we want to have fun talking about a shitpost we do kind of have to decide on an assumptive position on the meme that can’t talk back.
People making assumptions is the issue.
There’s assumptions involved in detecting satire from just text as well. You would just have a Reverse Poe’s law where “any extreme views can be mistaken by some readers for satire of those views without clear indicator of the author’s intent”.
Normally when people say or type things we (justifiably) assume that to be what they mean, which is why satire works much better when spoken because intonation can make the satire explicit without changing the words or saying it out loud.
As with most things, if you are competent, a degree doesn’t really matter. The degree is just a shortcut, and even if it’s checked it’s no guarantee you are otherwise competent. You’re expected to have picked up competency during the time you got your degree.
So this probably works if you are otherwise competent, but if you’re not it’s just going to lead to increased scrutiny (Because hey, you should know these things) and if someone does end up checking up on you it’s a great way to get fired with cause. Depending on how tight knit your industry is that can still make things very hard for you.
And of course, once this becomes frequent enough, you’d be surprised how quickly checking will become the norm again.
One likely reason the backlash has been so strong is because, on a platform as close-knit as Reddit, betrayal cuts deep.
Another laughable quote after the APIcalypse, at least for the people that remained on Reddit after being totally ok with being betrayed.
Yeah, it’s literally all over this thread, not exactly a secret. It’s kind of a weird nitpick of my comment, considering it’s just a way of phrasing things. If I give an alcoholic some money, I will say “they might use that to buy booze”. Because I am sure they buy booze, but they might use my money to buy some food instead. Not every single dollar you give the developers will go to ml.
You’re not required to do anything, let alone directly funding ml. That’s not what I am arguing for. I am arguing for you to support Lemmy despite the chance some of it might go to ml.
It goes the other way too, the developers probably disagree with a large part of the beliefs of people using lemmy, yet they also put in their time to create and foster it, which we never had to pay for either. They did it for the reasons they mention (free spaces, not owned by corporations that suck their users dry), which is separate from their other political positions.
I didn’t say AI would solve that, but I’ll re-iterate the point I’m making differently:
_
_
_