xAI Grok is a very powerful AI dialogue system that can generate smooth and natural replies in just a few seconds by inputting prompts. Grok stands out in the industry with its unique humorous style and real-time web information retrieval capabilities. Today, Grok has emerged in multiple innovative fields, and its influence is rapidly expanding. Whether for daily conversations, creative writing, or technical analysis and code debugging, Grok can provide insightful intelligent assistance, bringing a new dimension of support to users’ decision-making and creation. This document mainly introduces the usage process of the Grok Chat Completion API, allowing us to easily utilize the official Grok dialogue features.Documentation Index
Fetch the complete documentation index at: https://docs.acedata.cloud/llms.txt
Use this file to discover all available pages before exploring further.
Application Process
To use the Grok Chat Completion API, you can first visit the Grok Chat Completion API page and click the “Acquire” button to obtain the credentials needed for the request:
If you are not logged in or registered, you will be automatically redirected to the login page inviting you to register and log in. After logging in or registering, you will be automatically returned to the current page.
Upon the first application, there will be a free quota provided, allowing you to use the API for free.
Basic Usage
Next, you can fill in the corresponding content on the interface, as shown in the figure:
authorization, which can be selected directly from the dropdown list. The other parameter is model, which is the category of the Grok official model we choose to use. Here we mainly have 8 types of models; details can be found in the models we provide. The last parameter is messages, which is an array of our input questions. It is an array that allows multiple questions to be uploaded simultaneously, with each question containing role and content. The role indicates the role of the questioner, and we provide three identities: user, assistant, and system. The other content is the specific content of our question.
You can also notice that there is corresponding code generation on the right side; you can copy the code to run directly or click the “Try” button for testing.
Common optional parameters:
max_tokens: Limits the maximum number of tokens for a single reply.temperature: Generates randomness, between 0-2, with larger values being more divergent.n: How many candidate replies to generate at once.

id: The ID generated for this dialogue task, used to uniquely identify this dialogue task.model: The selected Grok official model.choices: The response information provided by Grok for the question.usage: Statistical information regarding the tokens for this Q&A.
choices contains the response information from Grok, and the choices inside it shows the specific information of Grok’s response, as can be seen in the figure.

content field inside choices contains the specific content of Grok’s reply.
Streaming Response
This interface also supports streaming responses, which is very useful for web integration, allowing the webpage to achieve a word-by-word display effect. If you want to return responses in a streaming manner, you can change thestream parameter in the request header to true.
Modify as shown in the figure, but the calling code needs to have corresponding changes to support streaming responses.

stream to true, the API will return the corresponding JSON data line by line, and we need to make corresponding modifications at the code level to obtain the line-by-line results.
Python sample calling code:
data in the response, and the choices in data is the latest response content, consistent with the content introduced above. The choices is the newly added response content, and you can integrate it into your system based on the results. The end of the streaming response is determined by the content of data; if the content is [DONE], it indicates that the streaming response has completely ended. The returned data result has multiple fields, described as follows:
id, the ID generated for this conversation task, used to uniquely identify this conversation task.model, the selected Grok official model.choices, the response information provided by Grok for the question words.
Multi-turn Conversation
If you want to integrate multi-turn conversation functionality, you need to upload multiple question words in themessages field. The specific example of multiple question words is shown in the image below:

choices is consistent with the basic usage content, which includes the specific content of Grok’s responses to multiple conversations, allowing you to answer corresponding questions based on multiple conversation contents.
Error Handling
When calling the API, if an error occurs, the API will return the corresponding error code and message. For example:400 token_mismatched: Bad request, possibly due to missing or invalid parameters.400 api_not_implemented: Bad request, possibly due to missing or invalid parameters.401 invalid_token: Unauthorized, invalid or missing authorization token.429 too_many_requests: Too many requests, you have exceeded the rate limit.500 api_error: Internal server error, something went wrong on the server.

