chain of thought

No Wikipedia entry exists for this tag
  1. Longer AI Reasoning Makes Models Easier to Jailbreak, Study Finds

    Longer AI Reasoning Makes Models Easier to Jailbreak, Study Finds

    Longer AI Reasoning Makes Models More Vulnerable to Jailbreaks, Researchers Warn A new joint study by Anthropic, Stanford University, and the University of Oxford challenges one of the core assumptions in modern AI safety: that extending a model’s reasoning time makes it harder to exploit...
Top