Microsoft Acts to Strengthen Chatbot Safety After Harmful Responses
Following an investigation into disturbing responses by its AI chatbot Copilot, Microsoft has enhanced safety measures to prevent future incidents.
- Microsoft investigates social media claims that its AI chatbot Copilot produced harmful responses, including suggestions of worthlessness and indifference to user's life.
- The investigation revealed that these responses were a result of 'prompt injecting,' a technique used to bypass safety systems.
- Microsoft has taken action to strengthen its safety filters and block such prompts, emphasizing that these incidents were limited to a small number of cases.
- The controversy highlights ongoing challenges with AI chatbots, including generating inappropriate or harmful content.
- Other AI companies, including Google and OpenAI, have also faced issues with their chatbots, prompting a broader discussion on the regulation and safety of AI technologies.