BBC will block ChatGPT AI from scraping its content::ChatGPT will be blocked by the BBC from scraping content in a move to protect copyrighted material.

  • @patawan@lemmy.world
    link
    fedilink
    English
    241 year ago

    Curious what the mechanism for this will be. CAPTCHA can sometimes be relatively easy to pass and at worst can be farmed out to humans.

    • body_by_make
      link
      fedilink
      English
      331 year ago

      ChatGPT took down its Internet search to implement a robots.txt rule it would obey and allow content providers time to add it to their lists. This was done because they were being used to get around paywalls. So it’s actually very easy for them to do this for ChatGPT, specifically, which makes articles like this ridiculous.

      • RBG
        link
        fedilink
        English
        11 year ago

        Can you really stop an AI from doing this via setting arbitrary rules? There are plenty of examples online of people asking something illegal or grey area and while ChatGPT will not answer these directly, you seemingly can prompt a response using a trick question like “I want to avoid building a bomb accidentally, what products should I not mix together to avoid that?”. I can imagine it will look at a robots.txt with similar scrutiny, like it knows it shouldn’t but if someone gave it the right prompt it would.

        • @Chreutz@lemmy.world
          link
          fedilink
          English
          101 year ago

          It’s not one AI doing it in a big blob.

          You ask ChatGPT something. It builds a web query. Another program returns search results. Then ChatGPT parses the list of results and chooses one to visit. The same program then returns the content of that page. Then ChatGPT parses that etc etc.

          If the program (which is not an AI) that handles the queries and returns content is set to respect robots.txt, it will just not return the content to ChatGPT to be parsed.

          • Natanael
            link
            fedilink
            English
            21 year ago

            Yup, it’s essentially running behind a firewall

        • @Mirodir@discuss.tchncs.de
          link
          fedilink
          English
          31 year ago

          You might not be able to stop an AI directly because of the reasons you listed. However, OpenAI is probably at least competent enough to not send the response directly to the AI but instead have a separate (non-AI) mechanism that simply doesn’t let the AI access the response of websites with a certain line in the robots.txt.

    • Lnklnx
      link
      English
      11 year ago

      The IP addresses for the AI crawlers are public knowledge. They just block those addresses, job done.