OpenAI researchers have published preliminary findings suggesting that increasing the computational resources allocated to reasoning models during inference can enhance their robustness against adversarial attacks. Their experiments with models like o1-preview and o1-mini demonstrated that providing more 'thinking' time often reduces the success rate of various attack methods, including those targeting mathematical tasks and factuality benchmarks. While this approach shows promise, the paper also notes significant exceptions and explores new attack vectors specifically designed for reasoning models. AI
Summary written by gemini-2.5-flash-lite from 1 source. How we write summaries →
RANK_REASON The cluster is based on a research paper from OpenAI detailing new findings on model robustness.