Apparently, stealing other people’s work to create product for money is now “fair use” as according to OpenAI because they are “innovating” (stealing). Yeah. Move fast and break things, huh?

“Because copyright today covers virtually every sort of human expression—including blogposts, photographs, forum posts, scraps of software code, and government documents—it would be impossible to train today’s leading AI models without using copyrighted materials,” wrote OpenAI in the House of Lords submission.

OpenAI claimed that the authors in that lawsuit “misconceive[d] the scope of copyright, failing to take into account the limitations and exceptions (including fair use) that properly leave room for innovations like the large language models now at the forefront of artificial intelligence.”

  • @intensely_human@lemm.ee
    link
    fedilink
    11 year ago

    I call it AI because it’s artificial and it’s intelligent. It’s not that complicated.

    The thing we have to remember is how scary and disruptive AI is. Given that fear, it is scary to acknowledge that we have AI emerging into our world. Because it is scary, that pushes us to want to ignore it.

    It’s called denial, and it’s the best explanation for why people aren’t willing to acknowledge that LLMs are AI.

    • @vexikron@lemmy.zip
      link
      fedilink
      41 year ago

      It meets almost none of the conceptions of intelligence at all.

      It is not capable of abstraction.

      It is capable of brute force understanding similarities between various images and text, and then presenting a wide array of text and images containing elements that reasonably well emulate a wide array of descriptors.

      This is convincing to many people that it has a large knowledge set.

      But that is not abstraction.

      It is not capable of logic.

      It is only capable of again brute force analyzing an astounding amount of content and then producing essentially the consensus view on answers to common logical problems.

      Ask it any complex logical question that has never been answered on the internet before and it will output irrelevant or inaccurate nonsense, likely just finding an answer to a similar but not identical question.

      The same goes for reasoning, planning, critical thinking and problem solving.

      If you ask it to do any of these things in a highly specific situation even giving it as much information as possible, if your situation is novel or even simply too complex, it will again just spit out a non sense answer that is basically going to be very inadequate and faulty because it will just draw elements together from the closest things it has been trained on, nearly certainly being contradictory or entirely dubious due to being unable to account for a particularly uncommon constraint, or constraints that are very uncommonly faced simultaneously.

      It is not creative, in the sense of being able to generate something novel or new.

      All it does is plagiarize elements of things that are popular and have many examples of and then attempt mix them together, but it will never generate a new art style or a new genre of music.

      It does not even really infer things, is not really capable of inference.

      It simply has a massive, astounding data set, and the ability to synthesize elements from this in a convincing way.

      In conclusion, you have no idea what you are talking about, and you yourself literally are one of the people who has failed the reverse Turing Test, likely because you are not very well very versed in the technicals of how this stuff actually works, thus proving my point that you simply believe it is AI because of its branding, with no critical thought applied whatsoever.

    • @ParsnipWitch@feddit.de
      link
      fedilink
      11 year ago

      Current models aren’t intelligent. Not even by the flimsy and unprecise definition of intelligence we currently have.

      Wanted to post a whole rant but then saw vexikron already did so I spare you xD