Microsoft’s AI security team reveals how hidden training backdoors quietly survive inside enterprise language models


  • Microsoft launches scanner to detect poisoned language models before deployment
  • Backdoor LLMs can hide malicious behavior until specific trigger phrases appear
  • The scanner identifies abnormal attention patterns linked to hidden backdoor triggers

Microsoft has announced the development of a new scanner designed to detect hidden backdoors in large, open-source language models used across enterprise environments.

The company says its tool aims to identify cases of model poisoning, a form of manipulation where malicious behavior is embedded directly into model weights during training.

Leave a Comment

Your email address will not be published. Required fields are marked *

Scroll to Top