Claude was being judgy, so I called it out. It immediately caved. Is verbal abuse a valid method of circumventing LLM censorship??

  • V0ldek@awful.systems
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    1 day ago

    Dunno, I disagree. It’s quite impossible for me to put myself in the shoes of a person who wouldn’t see a difference between shouting at an INANIMATE FUCKIN’ OBJECT vs at an actual person. As if saying “fuck off” to ChatGPT made me somehow more likely to then say “fuck off” to a waiter in a restaurant? That’s sociopath shit. If you need to “built the habit of being respectful” you have some deeper issues that should be solved by therapy, not by being nice to autocomplete.

    I’m a programmer since forever, I spend roughly 4h every day verbally abusing the C++ compiler because it’s godawful and can suck my balls. Doesn’t make me any more likely to then go to my colleague and verbally abuse them since, you know, they’re an actual person and I have empathy for them. If anything it’s therapeutic for me since I can vent some of my anger at a thing that doesn’t care. It’s like an equivalent of shouting into a pillow.

    • YourNetworkIsHaunted@awful.systems
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 day ago

      See, I feel like the one thing that Generative AI has been able to do consistently is to fool even some otherwise-reasonable people into thinking that there’s something like a person they’re talking to. One of the most toxic impacts that it’s had on online discourse and human-computer interactions in general is by introducing ambiguity into whether there’s a person on the other end of the line. On one hand, we need to wonder whether other posters on even this forum will Disregard All Previous Instructions. On the other hand, it’s a known fact that a lot of these “AI” tools are making heavy use of AGI technologies - A Guy in India. Before the bubble properly picked up my wife got contracted to work for a company that claimed to offer an AI personal assistant. Her job would have literally been to be the customer’s remote-working personal assistant. I like to think that her report to the regulators may have been part of what inspired these grifts to look internationally for their exploitable labor. I don’t think I need to get into the more recent examples here of all forums.

      Obviously yelling at your compiler isn’t going to lead to being an asshole to actual people any more than smashing a keyboard or cursing after missing a nail with a hammer. And to be fair most of the posters here (other than the drive-thrus) aren’t exactly lacking in class consciousness or human decency or whatever you want to call it, so I’m probably preaching to the choir. But I do think there’s a risk that injecting that ambiguity into the incidental relations we have with other people through our technologies (e.g. the chat window with tech support that could be a bot or a real agent depending on the stage of the conversation) is going to degrade the working conditions for a lot of real people, and the best way to avoid that is to set the norm that it’s better to be polite to the robot if it’s going to pretend to be a person.

      • Amoeba_Girl@awful.systems
        link
        fedilink
        English
        arrow-up
        3
        ·
        3 hours ago

        I agree with you, but i have to note the one thing that makes me lash out at an LLM chatbot is when the blighted thing pretends it’s anything at all like a person and capable of actual thought.

        • froztbyte@awful.systems
          link
          fedilink
          English
          arrow-up
          2
          ·
          3 hours ago

          is when the blighted thing pretends it’s anything at all like a person and …

          one of the helpful things that I think is worth reminding people of is that the thing described here is entirely the result of overt intent and choice on the part of humans shaping the fucking thing

          it isn’t an emergent property

          from the curation of data going into training to shaping “preferred output” (by tuning whatever properties my apply in xyz llm execution pipeline) this shit is all Super Polite by fucking design

          exact same shit applies for the prompts responding “oh yeah I’m totes alive! look, I feel doubted that you even asked! :sad:” - that shit is curated and engineered

          but I do hear where you’re coming from on the anger. 100%.