Anthropic Unveils 'Constitutional Classifiers' to Boost AI Safety!
Anthropic has rolled out its latest AI safety feature, the 'Constitutional Classifiers,' aimed at dramatically reducing jailbreak attempts in Claude AI. Targeting critical CBRN-related queries, this system minimizes successful jailbreaks from 86% to 4.4%. All this with minimal impact on legitimate queries and a slight increase in computational costs, paving the way for a safer AI future.
Feb 5