New Study Exposes AI Model Poisoning Risks
Sleeper agent-style backdoors hidden inside large language models are emerging as a serious and largely undetectable AI security threat, according to new research involving Microsoft’s AI red team. Researchers warn attackers can poison a model during training, embedding hidden triggers that activate malicious behavior on command. The threat centers on manipulated model weights. Attackers insert […]
