• Naz@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      1
      ·
      5 months ago

      Update: I’ve tried the expert topics and gaslighting and the model was able to give expert level information but would always correct itself, if given new information, even though it seemed absurd.

      However, the model would resist gas lighting for very well-known topics, such as claiming to be the “President of Mars”, it gave its logic for why the claim is false and was resistant to further attempts to try to convince it that this was true.

      Overall, this was a good experiment in doing real world testing on a large language model.

      Thanks for your suggestions – this is a problem that could be solved with future iterations of large language models! 💖