Grok 4.20
A fast, tool-focused variant of Grok 4.20 Beta with reasoning disabled for lower-latency agentic workloads and a massive 2M-token context window.
Fast agentic tool-calling with 2M-token context
Grok 4.20 is a text generation model developed by xAI, the AI division of X. This variant is specifically configured with reasoning disabled, meaning it skips the extended chain-of-thought process to deliver faster, lower-latency responses while still operating on the full Grok 4.20 architecture. It supports a context window of up to 2 million tokens, allowing it to ingest very long documents, large codebases, or extended conversation histories in a single pass. The model was made available via API in March 2026 as part of the Grok 4.20 Beta family, which also includes reasoning-enabled and multi-agent-tuned variants.
This model is designed for agentic and tool-centric workflows where response speed is a priority over deep step-by-step reasoning. It is well-suited for automated pipelines, coding agents, data-processing tasks, and any application where the model needs to call external tools rapidly and reliably. Its instruction-following behavior is tuned for consistency, making outputs predictable across repeated or templated prompts. Developers building low-latency AI systems or integrating LLM capabilities into production pipelines are the primary intended audience.
What Grok 4.20 supports
Massive Context Window
Processes up to 2 million tokens in a single pass, enabling ingestion of entire codebases, lengthy documents, or extended conversation histories without truncation.
Agentic Tool Calling
Optimized for rapid and reliable external tool invocation, making it suitable for automated agent frameworks and multi-step pipelines.
Non-Reasoning Mode
Reasoning is disabled by design, reducing latency by skipping extended chain-of-thought processing while retaining the underlying model's generation capabilities.
Instruction Following
Tuned for strong prompt adherence, producing consistent and predictable outputs across templated or repeated instructions.
Multimodal Input
Accepts input types beyond plain text, supporting diverse real-world task formats within a single model interface.
Text Generation
Generates coherent, contextually grounded text responses across a wide range of domains including coding, data processing, and conversational tasks.
Ready to build with Grok 4.20?
Get Started FreeBenchmark scores
Scores represent accuracy — the percentage of questions answered correctly on each test.
| Benchmark | What it tests | Score |
|---|---|---|
| GPQA Diamond | PhD-level science questions (biology, physics, chemistry) | 88.5% |
| HLE | Questions that challenge frontier models across many domains | 30.0% |
| SciCode | Scientific research coding and numerical methods | 44.7% |
Common questions about Grok 4.20
What is the context window size for Grok 4.20?
Grok 4.20 supports a context window of up to 2 million tokens, allowing it to process very long inputs in a single request.
Why is reasoning disabled in this variant?
Reasoning is disabled to reduce response latency. This makes the model faster and more suitable for agentic or tool-calling workflows where speed is prioritized over extended step-by-step reasoning.
What is the training data cutoff for Grok 4.20?
According to the model metadata, the training date is listed as March 2026.
Who publishes Grok 4.20?
Grok 4.20 is published by xAI, the AI division associated with X (formerly Twitter).
What types of workloads is this model best suited for?
This model is best suited for low-latency agentic systems such as automated assistants, coding agents, and data-processing pipelines where fast tool-calling and instruction adherence are more important than deep reasoning.
Is Grok 4.20 available via API?
Yes, Grok 4.20 Beta models were released via API in March 2026, as reflected in the model's dateAdded metadata.
What people think about Grok 4.20
Community discussion around Grok 4.20 on Reddit has been generally anticipatory and positive, with threads tracking its development from pre-training completion through its eventual API release. Users in the r/singularity community noted milestones such as Elon Musk confirming pre-training completion and the subsequent beta API launch in March 2026.
Some threads drew comparisons to other models releasing around the same period, reflecting broader interest in the competitive landscape rather than specific concerns about Grok 4.20 itself. The API release thread had modest engagement, suggesting early-adopter interest primarily among developers rather than widespread general discussion.
xAI to launch Grok 4.20 by Christmas
xAI Releases Grok 4.20 Beta Models via API
GLM-5 Coming in February! It's confirmed.
Elon Musk says that Grok 4.20 has completed its pre-training
Documentation & links
Parameters & options
Explore similar models
Start building with Grok 4.20
No API keys required. Create AI-powered workflows with Grok 4.20 in minutes — free.