XAI_API_KEY in your environment and point the base URL to https://api.x.ai/v1 — OpenClaw does not auto-detect this provider without both values.Grok gives your agents something they cannot get from Claude, GPT-4o, or Gemini: a training dataset that includes recent X platform content plus a knowledge cutoff that is weeks, not months, behind real time. Builders who connected Grok through OpenClaw in December 2024 saw immediate wins on trend-aware tasks that had previously required expensive custom scraping pipelines. That is the upside. The downside is a configuration that has two non-obvious steps most people skip, leading to silent failures that look like network errors. This guide fixes that.
xAI API vs Grok API: The Naming Confusion Explained
The single most common question from new users is whether "xAI API" and "Grok API" refer to the same thing. They do — but the naming matters for how you configure OpenClaw.
xAI is the company. It is Elon Musk's AI research organization. Grok is the model family that xAI builds and serves. When you call the xAI API, you are calling the company's inference endpoint and selecting a Grok model within that request. Think of it exactly like calling the Anthropic API and selecting claude-sonnet-4-5 — same structural logic.
Why does this matter for OpenClaw? Because the provider field in openclaw.yaml must be set to xai, not grok. Setting it to grok will cause a provider lookup failure that surfaces as a misleading "unknown model provider" error — not a useful error message pointing to the naming mismatch.
provider: xai. In the model field, always use grok-2 or grok-2-mini. The documentation sometimes shows "Grok API" which refers to the xAI endpoint — same thing, different label.XAI_API_KEY Setup
Getting your key takes under three minutes. Go to console.x.ai, create an account or sign in, then navigate to the API Keys section. Generate a new key, copy it immediately — it will not be shown again.
Set the key in your environment. The variable name OpenClaw expects is exactly XAI_API_KEY. Do not use GROK_API_KEY or XAI_KEY — both will cause silent auth failures where OpenClaw initializes but every model call returns a 401.
# .env file
XAI_API_KEY=xai-your-key-here
# or export directly in terminal
export XAI_API_KEY="xai-your-key-here"
Then configure openclaw.yaml:
model:
provider: xai
model_id: grok-2
base_url: https://api.x.ai/v1
max_tokens: 8192
temperature: 0.7
The base_url field is required. OpenClaw does not have a hardcoded fallback for the xAI endpoint the way it does for OpenAI and Anthropic. If you omit it, the client will attempt the OpenAI default endpoint and fail with a 404 that looks like a network problem.
After configuring, run openclaw doctor to verify the connection before building anything else on top of it.
grok-2 vs grok-2-mini: Which One to Run
xAI currently serves two primary models through the API. Here is how they compare for agent workloads:
| Property | grok-2 | grok-2-mini |
|---|---|---|
| Context window | 131,072 tokens | 131,072 tokens |
| Relative cost | ~$5 / 1M input tokens | ~$0.60 / 1M input tokens |
| Response speed | ~60–80 tokens/sec | ~120–150 tokens/sec |
| Best for | Reasoning, long-form, analysis | Classification, routing, summaries |
| Tool calling | Yes | Yes (limited) |
| Vision input | Yes | No |
The context window is identical between the two models, which surprises most people. The meaningful differences are cost, speed, and reasoning depth. For agent orchestration pipelines where Grok handles a routing or classification step, grok-2-mini is the correct call. For the terminal reasoning step that produces a final output, grok-2 is worth the cost premium.
Real-Time X Data Access: The Unique Advantage
This is where Grok separates from every other provider in the OpenClaw ecosystem. The model's training data includes X platform posts, and the knowledge cutoff is updated far more frequently than competing models — sometimes within weeks of current events.
Early adopters who built social listening agents reported that Grok could answer questions like "What is the current sentiment on X around [topic]?" with significantly higher accuracy than Claude or GPT-4o, without requiring any external API calls to retrieve X data. The model has absorbed the platform natively.
Web search capability is available through the Grok API as a tool, distinct from the base model knowledge. When you enable it in your OpenClaw agent config, the model can execute live web searches and inject results into its context. This is the combination that makes Grok genuinely differentiated for current-events agents.
What Early Adopters Found Surprising
The community who integrated Grok in the weeks following the API launch surfaced three consistent surprises.
Speed was faster than expected. Most builders expected Grok to be in the same tier as Claude or GPT-4o. In practice, response latency for grok-2 averaged lower than GPT-4 Turbo for similar task types — closer to 2–4 seconds for medium-length outputs rather than 5–8 seconds.
Real-time data usefulness was higher than the spec implied. Builders expecting vague awareness of recent events instead found Grok handling specific questions about content posted weeks before the query with confident, accurate answers. This drove immediate adoption for news summarization and trend analysis agents.
Cost was competitive. At roughly $5 per million input tokens for grok-2, the pricing sits below GPT-4 Turbo and in the same range as Claude 3.5 Sonnet. For agents that run many short queries, grok-2-mini at $0.60 per million tokens is one of the cheapest capable models available through OpenClaw.
Limitations to Know Before You Build
No provider is the right choice for every task. Here is what Grok cannot do as of early 2025.
- No image generation. Grok can receive and analyze images (with grok-2), but it cannot produce them. If your agent pipeline requires image synthesis, Grok is not a standalone solution.
- No fine-tuning endpoint. xAI does not currently expose a fine-tuning API. You cannot train a custom Grok variant on your domain data.
- Smaller tool library. The Grok API supports function calling, but the ecosystem of pre-built tool integrations in OpenClaw's tool registry is smaller for xAI than for Anthropic or OpenAI.
- Rate limits on new accounts. Free and new paid accounts start with conservative rate limits. Plan for this in production — scale your account tier before going live.
- Context window at 131K. Sufficient for most tasks, but if you are processing multi-hundred-page documents in a single call, you need MiniMax (1M context) or Gemini instead.
When to Choose Grok Over Claude for Agent Tasks
Claude is the default choice in most OpenClaw setups for good reason: it has the strongest reasoning, the largest tool ecosystem, and the most extensive safety tuning. But "default choice" is not the same as "best choice for every task."
Choose Grok when your agent needs any of the following:
- Awareness of events from the past 30–90 days without requiring live search
- X platform sentiment or trending topic context
- Less restrictive content generation for marketing or opinion-heavy copy
- Cost-sensitive pipelines where you need capable reasoning below Claude's price point
- Fast response time for interactive, chat-style agent interfaces
Stick with Claude when your agent needs deep document analysis, complex multi-step reasoning, or the widest possible tool support from OpenClaw's built-in registry.
Common Config Mistakes
Three mistakes account for roughly 80% of Grok integration failures in the community Discord.
Mistake 1: Wrong Model ID
Using grok-beta (the older preview model ID) instead of grok-2. The beta model is still technically available but returns degraded quality results and does not support vision input. Always use grok-2 or grok-2-mini.
Mistake 2: Missing base_url
Setting provider: xai but omitting the base_url field. This causes OpenClaw to fall back to the OpenAI endpoint, which rejects xAI auth tokens with a confusing 401 that looks like a bad API key rather than a wrong endpoint.
Mistake 3: Wrong Environment Variable Name
Using GROK_API_KEY instead of XAI_API_KEY. OpenClaw looks specifically for XAI_API_KEY. The wrong variable name means the key never gets loaded and every call returns an unauthenticated error.
# WRONG — these will not work
GROK_API_KEY=xai-...
XAI_KEY=xai-...
# CORRECT
XAI_API_KEY=xai-your-full-key-here
Run openclaw doctor --provider xai after setup. It checks all three of these in one command and tells you exactly which step failed.
Frequently Asked Questions
What is the difference between the xAI API and the Grok API?
xAI is the company; Grok is the model family. The xAI API is the endpoint you call, using your XAI_API_KEY, and it serves Grok models like grok-2 and grok-2-mini. In OpenClaw config, set provider: xai and choose your model in the model_id field. They are the same endpoint with different model selections — not two separate services.
How do I get an XAI_API_KEY for OpenClaw?
Sign up at console.x.ai, navigate to API Keys, and generate a key. Set it as XAI_API_KEY in your environment or .env file. OpenClaw picks it up automatically when provider is set to xai in openclaw.yaml. The key prefix will begin with xai- — this is correct.
Can Grok access real-time X (Twitter) data inside OpenClaw agents?
Grok's training includes X data and the model has built-in awareness of recent X trends, but real-time live post retrieval requires X API integration separately in your agent workflow. The model's knowledge window is significantly more recent than other providers, which makes it uniquely useful for recent-event queries without additional tooling.
What is the context window size for grok-2 in OpenClaw?
As of early 2025, grok-2 supports a 131,072 token context window. This is comparable to Claude 3.5 Sonnet but smaller than Gemini's 1M offering. For most agent workflows — document analysis, summarization, multi-turn conversations — 131K is sufficient without additional chunking logic.
When should I choose Grok over Claude for an OpenClaw agent?
Choose Grok when your agent needs awareness of recent events, X platform sentiment, or when Claude's response style is too conservative for your content use case. Grok is notably less restrictive on opinion-heavy or edgy content generation, which matters for marketing and social media agents.
Does Grok support image generation inside OpenClaw?
No. As of early 2025, xAI does not offer image generation through the Grok API. Grok can analyze images sent as input (grok-2 only) but cannot generate them. For image generation tasks inside an OpenClaw agent pipeline, use OpenAI's DALL-E or Stability AI as a separate tool call.
What are the most common config mistakes when setting up Grok in OpenClaw?
Three mistakes dominate: using the wrong model ID (grok-beta instead of grok-2), forgetting to set base_url to https://api.x.ai/v1, and using the wrong environment variable name (GROK_API_KEY instead of XAI_API_KEY). Run openclaw doctor --provider xai to catch all three at once.
Is grok-2-mini worth using over grok-2 for agent tasks?
grok-2-mini runs roughly 3x cheaper and responds faster. For classification, summarization, or routing tasks inside an agent pipeline where you need capable but not peak reasoning, mini is the better choice. Reserve grok-2 for the terminal reasoning step or tasks requiring vision input — that is where the cost premium pays off.
openclaw doctor --provider xai — it validates your key, endpoint, and model ID in under 10 seconds. No account changes needed.