Skip to content
Zhipu AI

Configure Zhipu AI

Zhipu AI is a leading cognitive intelligence large model company in China, providing the GLM-4 series of large language model services. GLM-4 supports ultra-long context, multimodal understanding, high-performance reasoning, and other features, widely used in dialogue, text generation, code assistance, and other scenarios.

1. Get Zhipu AI API Key

1.1 Access AI Open Platform

Visit the AI Open Platform and log in: https://open.bigmodel.cn/

Access AI Platform

1.2 Go to API Management Page

Click the user avatar in the upper right corner, select API Management.

Go to API Management Page

1.3 Create a New API Key

Click the Create API Key button.

Click Create Button

1.4 Set API Key Information

In the popup dialog:

  1. Enter a name for the API Key (e.g., CueMate)
  2. Select applicable models
  3. Click the Confirm button

Set API Key Information

1.5 Copy API Key

After successful creation, the system will display the API Key.

Important: This is the only time you can see the complete API Key, please copy and save it securely immediately.

Copy API Key

Click the copy button, and the API Key will be copied to your clipboard.

2. Configure Zhipu AI Model in CueMate

2.1 Go to Model Settings Page

After logging into the CueMate system, click Model Settings in the dropdown menu in the upper right corner.

Go to Model Settings

2.2 Add New Model

Click the Add Model button in the upper right corner.

Click Add Model

2.3 Select Zhipu AI Provider

In the popup dialog:

  1. Provider Type: Select Zhipu AI
  2. Click to automatically proceed to the next step

Select Zhipu AI

2.4 Fill in Configuration Information

Fill in the following information on the configuration page:

Basic Configuration

  1. Model Name: Give this model configuration a name (e.g., Zhipu GLM-4 Plus)
  2. API URL: Keep the default https://open.bigmodel.cn/api/paas/v4 (OpenAI compatible format)
  3. API Key: Paste the Zhipu AI API Key you just copied
  4. Model Version: Select the model ID to use, common models include:
    • glm-4-plus: Most powerful model, suitable for complex reasoning and deep analysis (max output 8K)
    • glm-4-long: Ultra-long text processing, supports 1M context (max output 8K)
    • glm-4-air: Lightweight efficient version, fast response (max output 8K)
    • glm-4-airx: Ultra-fast version, ultra-low latency (max output 8K)
    • glm-4-flash: Lightning response, real-time dialogue (max output 8K)
    • glm-4: Standard version, balances performance and cost (max output 8K)
    • glm-4v: Multimodal model, supports image understanding (max output 8K)
    • glm-4v-plus: Multimodal enhanced version, supports 2-hour video, 4K images (max output 8K)
    • glm-3-turbo: Affordable version, daily dialogue (max output 4K)

Fill in Basic Configuration

Advanced Configuration (Optional)

Expand the Advanced Configuration panel to adjust the following parameters:

Parameters adjustable in CueMate interface:

  1. Temperature: Controls output randomness

    • Range: 0-1
    • Recommended Value: 0.7
    • Function: Higher values produce more random and creative output, lower values produce more stable and conservative output
    • Usage Suggestions:
      • Creative writing/brainstorming: 0.8-0.95
      • Regular conversation/Q&A: 0.6-0.8
      • Code generation/precise tasks: 0.3-0.5
    • Note: Zhipu AI's temperature range is 0-1, different from OpenAI's 0-2
  2. Max Tokens: Limits single output length

    • Range: 256 - 8192 (depending on model)
    • Recommended Value: 4096
    • Function: Controls the maximum word count per model response
    • Model Limits:
      • GLM-4 series: Max 8K tokens
      • GLM-3-turbo: Max 4K tokens
    • Usage Suggestions:
      • Brief Q&A: 1024-2048
      • Regular conversation: 4096-8192
      • Long text generation: 8192 (maximum)

Advanced Configuration

Other advanced parameters supported by Zhipu AI API:

While the CueMate interface only provides temperature and max_tokens adjustments, if you call Zhipu AI directly via API, you can also use the following advanced parameters (Zhipu AI uses OpenAI compatible API format):

  1. top_p (nucleus sampling)

    • Range: 0-1
    • Default: 0.7
    • Function: Samples from the smallest candidate set whose cumulative probability reaches p
    • Relationship with temperature: Can be used together
    • Usage Suggestions:
      • Maintain diversity: 0.7-0.95
      • More conservative output: 0.5-0.7
  2. do_sample

    • Type: Boolean
    • Default: true
    • Function: Enables random sampling (set to false for greedy decoding)
    • Use Cases:
      • Creative tasks: true (enable sampling)
      • Deterministic tasks: false (greedy decoding)
  3. stop (stop sequences)

    • Type: String or array
    • Default: null
    • Maximum: 4 strings
    • Function: Stops when generated content contains specified strings
    • Example: ["###", "User:", "\n\n"]
    • Use Cases:
      • Structured output: Use delimiters to control format
      • Dialogue systems: Prevent model from speaking for user
  4. stream (streaming output)

    • Type: Boolean
    • Default: false
    • Function: Enables SSE streaming return, returns as it generates
    • In CueMate: Automatically handled, no manual setting needed
  5. tools (tool calling)

    • Type: Object array
    • Function: Defines tools/functions the model can call
    • Use Cases: Function Calling, Agent applications
    • Example:
      json
      {
        "tools": [
          {
            "type": "function",
            "function": {
              "name": "get_weather",
              "description": "Get weather for specified city",
              "parameters": {
                "type": "object",
                "properties": {
                  "city": {"type": "string"}
                }
              }
            }
          }
        ]
      }

Zhipu AI Special Parameters:

  1. request_id
    • Type: String
    • Function: User-provided unique ID for request tracking
    • Usage Suggestion: Pass unique identifier for tracking and debugging
No.Scenariotemperaturemax_tokenstop_pdo_samplestop
1Creative Writing0.8-0.954096-81920.9truenull
2Code Generation0.2-0.52048-40960.7truenull
3Q&A System0.6-0.81024-20480.7truenull
4Summarization0.3-0.5512-10240.7truenull
5Deterministic Tasks020481.0falsenull

2.5 Test Connection

After filling in the configuration, click the Test Connection button to verify if the configuration is correct.

Test Connection

If the configuration is correct, a success message will be displayed along with a sample response from the model.

Test Success

If the configuration is incorrect, an error log will be displayed, and you can view specific error information through log management.

2.6 Save Configuration

After successful testing, click the Save button to complete the model configuration.

Save Configuration

3. Use the Model

Go to the system settings page through the dropdown menu in the upper right corner, and select the model configuration you want to use in the LLM provider section.

After configuration, you can select this model in features like interview training and question generation. You can also select this model configuration for a specific interview in the interview options.

Select Model

4. Supported Model List

No.Model NameModel IDMax OutputUse Case
1GLM-4 Plusglm-4-plus8K tokensMost powerful version, complex reasoning
2GLM-4 Longglm-4-long8K tokensLong text processing, supports 1M context
3GLM-4 Airglm-4-air8K tokensLightweight, fast response
4GLM-4 AirXglm-4-airx8K tokensUltra-fast version, ultra-low latency
5GLM-4 Flashglm-4-flash8K tokensReal-time dialogue, lightning response
6GLM-4glm-48K tokensStandard version, technical interviews
7GLM-4Vglm-4v8K tokensMultimodal, supports image understanding
8GLM-4V Plusglm-4v-plus8K tokensMultimodal enhanced, supports 2-hour video
9GLM-3 Turboglm-3-turbo4K tokensAffordable, regular dialogue

5. FAQ

5.1 Invalid API Key

Symptom: API Key error when testing connection

Solutions:

  1. Check if API Key is completely copied
  2. Confirm API Key has not expired or been disabled
  3. Check if account has available credits

5.2 Request Timeout

Symptom: Long wait time with no response when testing connection or using the model

Solutions:

  1. Check if network connection is normal
  2. Check firewall settings
  3. Confirm Zhipu AI service status is normal

5.3 Insufficient Quota

Symptom: Quota exhausted message

Solutions:

  1. Log in to Zhipu AI platform to check account balance
  2. Top up or request more quota
  3. Optimize usage frequency

Released under the GPL-3.0 License.