This might also be an automatic response to prevent discussion. Although I’m not sure since it’s MS’ AI.

  • otp@sh.itjust.works
    link
    fedilink
    arrow-up
    0
    ·
    edit-2
    2 years ago

    I think the LLM won here. If you’re being accusational and outright saying its previous statement is a lie, you’ve already made up your mind. The chatbot knows it can’t change your mind, so it suggests changing the topic.

    It’s not a spokesperson/bot for Microsoft, not a lawyer. So it knows when it should shut itself off.

  • eveninghere@beehaw.org
    link
    fedilink
    arrow-up
    0
    ·
    edit-2
    2 years ago

    This is actually an unfair experiment. This behavior is not specific to questions about MS. Copilot is simply incapable of this type of discussion.

    Copilot tends to just paraphrase text it read, and when I challenge the content, it ends the conversation like this, instead of engaging in a meaningful dialogue.