• 𝙣𝙪𝙠𝙚@yah.lol
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 year ago

    Why did I do this?

    Because I don’t want my material used for training LLM’s.

    Good luck with that. Sure, maybe OpenAI respects your robots.txt and doesn’t scan you, but I seriously doubt everyone is going to abide by the honor system here.

    • Lvxferre
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      1
      ·
      1 year ago

      I seriously doubt everyone is going to abide by the honor system here.

      I think that we (people in general) should be creating honeytraps to punish the ones not abiding to the honour system. For example:

      • create pages with text composed of random babble
      • link those pages in the site, somewhere that humans won’t access
      • exclude those pages from bot crawling in robots.txt

      Bots abiding to the honour system will ignore those pages and move on. The ones not abiding to it will crawl through those pages and feed them into their owners’ AI models, that will be worse in result.

  • Lvxferre
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    1
    ·
    edit-2
    1 year ago

    But I don’t want to have to dispel wrong notions because of a dumb bot using statistics to pretend to speak.

    Perfect conclusion to a piece of text addressing the main issue with practical usage of LLMs: they get things wrong so often that you can’t rely on them, leading you to a state that is worse than a simple “I don’t know” - the “I believe that I know, but actually I don’t know”.


    From the comments in Techddit Hacker News. Content warning: I’ll quote rampant stupidity. Specially given HN worships this sort of bot.

    [You should take down the documentation entirely, if you want to prevent incorrect interpretations of things.]¹ [The LLMs won’t be the ones emailing you]², [the people who would get things wrong if the LLM provided some kind of confident wrong answer would probably simply not read your documentation as the vast majority of users do not]³.

    1. That’s a bloody strawman. Howard’s claim is not that he wants to get rid of all incorrect interpretation; it boils down to GPT increasing his support workload, instead of decreasing it.
    2. No shit Sherlock.
    3. False dichotomy. Here’s a third possibility: user asks ChatGPT about the software, ChatGPT feeds the user wrong info, user realises “it doesn’t fucking work!” and contacts dev, dev tells user to RTFD, user does it while contextualising it around the wrong info that ChatGPT fed him, issue doesn’t go away, user keeps nagging dev.

    [This is such a technopurist take. People who use LLM’s already know they can give wrong information.]¹ [Your documentation won’t be able to cover every single possible contextual scenario that an LLM [SIC - a bot using that LLM] can help with.]²

    1. They do. But knowing what is wrong often depends on knowing what is correct. And if they were informed enough to say “this piece of info is wrong, this one is right…”, they wouldn’t be asking the bot on first place.
    2. If a source of info won’t help you, odds are that the bot won’t help you either, even after being fed that source of info.

    I do think that chatbots can be useful in some situations, mind you; just like my tarot deck. They don’t give you real answers, but they give you ideas and things to think about.