Microsoft said the bug meant that its Copilot AI chatbot was reading and summarizing paying customers' confidential emails, bypassing data-protection policies.
Right, the real issue is that there needs to be a layer between the app and the LLM which handles authorization and decides whether the data is confidential before it’s ever sent to a remote server. It’s not even an LLM issue, it’s just bad architecture in general.
Right, the real issue is that there needs to be a layer between the app and the LLM which handles authorization and decides whether the data is confidential before it’s ever sent to a remote server. It’s not even an LLM issue, it’s just bad architecture in general.