This might also be an automatic response to prevent discussion. Although I’m not sure since it’s MS’ AI.

    • Hotzilla@sopuli.xyz
      link
      fedilink
      arrow-up
      0
      ·
      2 years ago

      Copilot runs with GPT4-turbo. It is not trained differently than openai’s GPT4-turbo, but it has different system prompts than openai, which tend to make it more easy to just quit discussion. I have never seen openai to say that I will stop this conversation, but copilot does it daily.

  • eveninghere@beehaw.org
    link
    fedilink
    arrow-up
    0
    ·
    edit-2
    2 years ago

    This is actually an unfair experiment. This behavior is not specific to questions about MS. Copilot is simply incapable of this type of discussion.

    Copilot tends to just paraphrase text it read, and when I challenge the content, it ends the conversation like this, instead of engaging in a meaningful dialogue.

  • otp@sh.itjust.works
    link
    fedilink
    arrow-up
    0
    ·
    edit-2
    2 years ago

    I think the LLM won here. If you’re being accusational and outright saying its previous statement is a lie, you’ve already made up your mind. The chatbot knows it can’t change your mind, so it suggests changing the topic.

    It’s not a spokesperson/bot for Microsoft, not a lawyer. So it knows when it should shut itself off.