Blake is a researcher on the AI Red Team at Microsoft, where he studies safety and security vulnerabilities in generative AI systems. His work focuses on understanding and stress-testing model behavior under adversarial conditions, with the goal of mitigating near-term risks while informing the design of more robust AI systems in the long term.
We’re releasing new research on detecting backdoors in open-weight language models and highlighting a practical scanner designed to detect backdoored models at scale and improve overall trust in AI systems.