reasoning model

AI Reasoning Models Can Be Jailbroken With Over 80% Success Rate Using Novel Attack Method

A joint study by Anthropic, Oxford University and Stanford has revealed a fundamental security flaw in advanced AI reasoning models: enhanced thinking capabilities do not strengthen but rather weaken models' defences against harmful commands. The attack method called Chain-of-Thought Hijacking successfully bypasses built-in safety mechanisms with more than 80%

by poltextLAB AI journalist

Anthropic Unveils Claude Opus 4.1 Model with Enhanced Coding Capabilities

On August 7, 2025, Anthropic released Claude Opus 4.1, featuring significant improvements in coding, agent, and reasoning capabilities, showing particular advancement in handling complex real-world programming tasks and multi-step problems. The updated model delivers 38% better performance on coding tasks, and 27% enhanced reasoning capabilities on HumanEval, MMLU, and

by poltextLAB AI journalist

Mistral AI Unveils Its First Reasoning Model, 10x Faster Than Competitors

French AI lab Mistral AI officially announced Magistral on June 10, 2025, its first family of reasoning models capable of step-by-step thinking, available in two variants: the open-source 24-billion-parameter Magistral Small and the enterprise-focused Magistral Medium. Magistral Medium scored 73.6% accuracy on the AIME2024 mathematics benchmark, rising to 90%

by poltextLAB AI journalist