xAI Grok 4
The xai.grok-4 model has better performance than its predecessor, Grok 3, and excels at enterprise use cases such as data extraction, coding, and summarizing text. This model has a deep domain knowledge in finance, healthcare, law, and science.
Available in These Regions
- US East (Ashburn) (on-demand only)
 - US Midwest (Chicago) (on-demand only)
 - US West (Phoenix) (on-demand only)
 
External Calls
The xAI Grok models are hosted in an OCI data center, in a tenancy provisioned for xAI. The xAI Grok models, which can be accessed through the OCI Generative AI service, are managed by xAI.
Access this Model
Key Features
- Model name in OCI
Generative AI:
xai.grok-4 - Available On-Demand: Access this model on-demand, through the Console playground or the API.
 - Multimodal support: Input text and images and get a text output.
 - Knowledge: Has a deep domain knowledge in finance, healthcare, law, and science.
 - Context Length: 128,000 tokens (maximum prompt + response length is 128,000 tokens for keeping the context). In the playground, the response length is capped at 16,000 tokens for each run, but the context remains 128,000 tokens.
 - Excels at These Use Cases: Data extraction, coding, and summarizing text
 - Function Calling: Yes, through the API.
 - Structured Outputs: Yes.
 - Has Reasoning: Yes. For reasoning problems increase the maximum output tokens. See Model Parameters.
 - 
Cached Input Tokens: Yes
- Token count: See the 
cachedTokensattribute in the PromptTokensDetails Reference API. - Pricing: See the Pricing Page.
 
Important note: The cached‑input feature is available in both the playground and the API. However, that information can only be retrieved through the API.
 - Token count: See the 
 - Knowledge Cutoff: November 2024
 
Limits
- Image Inputs
 - 
- Console: Upload one or more 
.pngor.jpgimages, each 5 MB or smaller. - API: Submit a 
base64encoded version of an image, ensuring that each converted image is more than 512 and less than 1,792 tokens. For example, a 512 x 512 image typically converts to around 1,610 tokens. 
 - Console: Upload one or more 
 
On-Demand Mode
The Grok models are available only in the on-demand mode.
| Model Name | OCI Model Name | Pricing Page Product Name | 
|---|---|---|
| xAI Grok 4 | xai.grok-4 | 
xAI – Grok 4 Prices are listed for: 
  | 
- 
You pay as you go for each inference call when you use the models in the playground or when you call the models through the API.
 - Low barrier to start using Generative AI.
 - Great for experimentation, proof of concept, and model evaluation.
 - Available for the pretrained models in regions not listed as (dedicated AI cluster only).
 
Release Date
| Model | General Availability Release Date | On-Demand Retirement Date | Dedicated Mode Retirement Date | 
|---|---|---|---|
xai.grok-4 | 
2025-07-23 | Tentative | This model isn't available for the dedicated mode. | 
Model Parameters
To change the model responses, you can change the values of the following parameters in the playground or the API.
- Maximum output tokens
 - 
The maximum number of tokens that you want the model to generate for each response. Estimate four characters per token. Because you're prompting a chat model, the response depends on the prompt and each response doesn't necessarily use up the maximum allocated tokens. The maximum prompt + output length is 128,000 tokens for each run.
Tip
For large inputs with difficult problems, set a high value for the maximum output tokens parameter. See Troubleshooting. - Temperature
 - 
The level of randomness used to generate the output text. Min: 0, Max: 2
Tip
Start with the temperature set to 0 or less than one, and increase the temperature as you regenerate the prompts for a more creative output. High temperatures can introduce hallucinations and factually incorrect information. - Top p
 - 
A sampling method that controls the cumulative probability of the top tokens to consider for the next token. Assign
pa decimal number between 0 and 1 for the probability. For example, enter 0.75 for the top 75 percent to be considered. Setpto 1 to consider all tokens. 
The xai.grok-4 model has reasoning, but doesn't support the reasoning_effort parameter used in the Grok 3 mini and Grok 3 mini fast models. If you specify the reasoning_effort parameter in the API for the xai.grok-4 model, you get an error response.
Troubleshooting
Issue: The Grok 4 model doesn't respond.
Cause: The Maximum output tokens parameter in the playground or the max_tokens parameter in the API is likely too low.
Action: Increase the maximum output tokens parameter.
Reason: For difficult problems that require reasoning and problem-solving, and for large sophisticated inputs, the xai.grok-4 model tends to think and consumes many tokens, so if the max_tokens parameter is too low, the model uses the allocated tokens and doesn't return a final response.