Anthropic Launches A New AI Model That ‘Thinks’ As Long As You Want

Anthropic has introduced Claude 3.7 Sonnet, a cutting-edge hybrid AI reasoning model that allows users to control how long the AI “thinks” before providing answers. Unlike traditional AI models, Claude 3.7 Sonnet integrates both real-time responses and deep reasoning capabilities within a single system.

The Claude 3.7 Sonnet AI model is part of Anthropic’s strategy to simplify the AI user experience by eliminating complex model selection. Users can activate Claude’s reasoning mode, enabling it to process information more thoroughly before responding. Eventually, Anthropic aims to make Claude autonomous in selecting reasoning time based on query complexity, improving artificial intelligence automation.

Premium users of Claude AI will gain access to its advanced reasoning features, while free-tier users receive a standard version. Anthropic states that Claude 3.7 Sonnet outperforms Claude 3.5 Sonnet, even without reasoning mode enabled. The model is priced at $3 per million input tokens and $15 per million output tokens, making it more expensive than OpenAI’s o3-mini or DeepSeek’s R1. However, unlike these models, Claude 3.7 Sonnet functions as both an instant-response and deep-reasoning model, combining AI-powered natural language processing with automated problem-solving.

Anthropic’s AI reasoning technology aligns with the industry’s focus on logical step-by-step problem-solving. Competitors like OpenAI (o3-mini), Google (Gemini 2.0 Flash Thinking), and xAI (Grok 3 Think) have also adopted similar techniques to enhance AI accuracy by breaking problems into logical steps. While these reasoning-based AI models do not truly “think” like humans, they employ deductive reasoning algorithms to improve AI-generated responses.

One of Claude 3.7 Sonnet’s standout features is its “visible scratch pad,” allowing users to observe the AI’s internal processing as it formulates answers. However, Anthropic notes that some reasoning steps may be redacted for trust and safety compliance. The model has also been optimized for real-world applications, such as complex coding challenges and API-based AI interactions.

In benchmark tests, Claude 3.7 Sonnet demonstrated 62.3% accuracy on SWE-Bench, a real-world coding task evaluation, outperforming OpenAI’s o3-mini (49.3%). Similarly, in TAU-Bench, which measures an AI’s ability to interact with simulated users and external APIs, Claude 3.7 Sonnet scored 81.2%, compared to OpenAI’s o1 model (73.5%).

Beyond reasoning, Anthropic has improved Claude’s refusal rate, reducing unnecessary denials of user prompts by 45% compared to previous models. This update comes as AI companies adjust content moderation policies, balancing AI safety protocols with broader accessibility for users.

Alongside Claude 3.7 Sonnet, Anthropic is launching Claude Code, an agentic AI-powered coding assistant. Available as a research preview, Claude Code integrates directly with developer environments, allowing AI-driven code analysis, modifications, testing, and GitHub integration through natural language commands. This tool will initially roll out to a limited number of AI developers on a first-come, first-served basis.

Leave a Reply

Your email address will not be published. Required fields are marked *