Skip to content

Model Settings

Model Settings is used to configure and manage all Large Language Models (LLMs), supporting both public cloud models (OpenAI, Claude, DeepSeek) and privately deployed models (Ollama, vLLM, Xinference). Through the unified configuration interface, you can add multiple models, manage credentials, test connection status, and configure model parameters.

1. Page Layout

1.1 Enter Model Settings Page

Click "Model Settings" in the top dropdown menu to enter the model management page.

Top Dropdown Menu

1.2 Layout Structure

The page uses a left-right split design:

Left Navigation Area

  • Tree structure displaying provider categories
  • Three-level classification: All Models, Public Models, Private Models
  • Each provider has a brand icon
  • Supports collapsing/expanding sidebar

Right Content Area

  • Top: Title, search box, refresh and add buttons
  • Main: Card-style model list, 2-column layout on desktop
  • Bottom: Pagination navigation

Model Card Information

  • Top left: Number, provider icon, and model name
  • Public/Private tag, connection status tag (Connected/Unavailable)
  • Provider, model type, base model, version, creator, creation time
  • Hover to show action buttons: Test Connection, Edit, Delete

Page Layout

2. Browse Models

2.1 Filter Using Navigation Tree

The left navigation tree organizes models in a tree structure:

All Models
├── Public Models
│   ├── OpenAI
│   ├── Claude
│   ├── DeepSeek
│   └── ...
└── Private Models
    ├── Ollama
    ├── vLLM
    └── ...

Operation Methods

  • Click "All Models" to show all models
  • Click "Public Models" or "Private Models" to show models in that category
  • Click a specific provider to show only that provider's models
  • Selected node is highlighted

Public vs Private

  • Public Models: Cloud services, data processed on provider's cloud (e.g., OpenAI, Claude)
  • Private Models: Local deployment, data processed on your device (e.g., Ollama, vLLM)

2.2 View Model List

Models are displayed as cards on the right, each card contains:

  • Number and provider icon
  • Model name and type tag
  • Connection status (green "Connected" / red "Unavailable")
  • Detailed info (provider, base model, version, creator, time)

Responsive Layout

  • Large screens: 2-column cards
  • Small screens: 1-column cards

Model List

2.3 Search and Filter

Search Function

  • Location: Top right search box
  • Search scope: Model name
  • Supports real-time search and clear

Filter Combination

  • Search and navigation filters can be used simultaneously
  • Example: Select "Public Models" then search "GPT", only shows public models containing "GPT"

Refresh List

  • Click "Refresh" button to get latest model status
  • Auto refresh: Automatically refreshes after adding, editing, or deleting models

Model Card

3. Add Model

3.1 Select Provider

Click the "Add Model" button in the top right corner to open the provider selection drawer.

Provider Selection Drawer

  • Width: 65% of page
  • Card grid: 2-column layout
  • Each card shows: Number, icon, provider name
  • Filter logic:
    • Select "Public Models" to show only public providers
    • Select "Private Models" to show only private providers
    • Select "All Models" to show all providers

Provider List Examples

Public Cloud Models (16):

  • International mainstream: OpenAI, Anthropic (Claude), Google Gemini, DeepSeek
  • Chinese major providers: Moonshot (Kimi), Zhipu AI, Tongyi Qianwen (Alibaba Cloud Bailian), Tencent Hunyuan, iFlytek Spark, Volcano Engine (Doubao), Baidu Qianfan, SenseTime SenseNova
  • Emerging platforms: Baichuan AI, MiniMax, StepFun, SiliconFlow

Cloud Platform Model Services (3):

  • Azure OpenAI, Amazon Bedrock, Tencent Cloud

Private Deployment Models (5):

  • Ollama, vLLM, Xinference, Local Model, Regolo

Select Provider

3.2 Configure Basic Information

After selecting a provider, enter the configuration interface (second-level drawer).

Basic Information

  1. Model Name (Required)
    • Maximum 64 characters
    • Examples: GPT-4 Turbo, Claude 3 Sonnet, DeepSeek Chat
    • Top right provides "View Configuration Guide" link, click to jump to the provider's configuration documentation

View Configuration Guide

At the top right of the model name input box, there's a "View Configuration Guide" link. Clicking will show a confirmation dialog, and after confirming, it will open the provider's configuration documentation in an external browser (client) or new tab (web).

View Configuration Guide

  1. Model Situation (Auto-selected)

    • Automatically selected based on provider type
    • Private Model: Local deployment, data processed on your device or environment
    • Public Model: Cloud service, data processed on provider's cloud
  2. Model Type (Required)

    • Current option: Large Language Model (llm)
  3. Base Model (Required)

    • Dropdown to select models supported by the provider
    • Supports search and custom input
    • Directly enter model name and press Enter to add
    • Examples:
      • OpenAI: gpt-5, gpt-5-mini, gpt-4.1, gpt-4o
      • Claude: claude-sonnet-4-5, claude-haiku-4-5, claude-opus-4-1
      • DeepSeek: deepseek-chat, deepseek-coder

Configure Basic Information

3.3 Configure Credentials

After selecting the base model, the credential configuration area is automatically displayed.

Common Credential Fields

  1. API Key (Password type)

    • Input box defaults to hidden display (• • • •)
    • Click eye icon to toggle show/hide
    • Placeholder: Shows format hint like "sk-xxx"
    • Required field, shows red border if empty
  2. Base URL (Text type, comes with default value, normally no modification needed)

  3. Other Fields

    • Depending on provider, may include: Organization ID, Project ID, Access Token, Secret Key
    • Each field has clear label and placeholder hints

Credential Configuration

3.4 Configure Advanced Parameters

Click the "Advanced Settings" tab to configure model runtime parameters (normally shows default values for the model, custom models have no defaults and need to be filled in).

Parameter Table

  • Table columns: Display Name, Parameter, Component Type, Default Value, Required, Actions
  • Top right "+ Add" button to add new parameters
  • Each row has "Delete" button on the right to delete the parameter

Parameter Configuration Items

  1. Display Name (label)

    • Name shown in the interface
    • Examples: Temperature, Max Tokens, Top P
  2. Parameter (param_key)

    • Actual parameter key passed to the model
    • Examples: temperature, max_tokens, top_p, frequency_penalty
    • Must match the model's API documentation
  3. Component Type (ui_type)

    • Options: Input box (input), Slider (slider), Switch (switch), Dropdown (select)
    • Determines how the parameter is displayed in conversation settings
  4. Default Value (value)

    • Default value for the parameter
    • Examples: temperature: 0.7, max_tokens: 2048
  5. Required (required)

    • Toggle control, on means required field

Common Parameter Examples

Display Name: Temperature
Parameter: temperature
Component Type: Slider
Default Value: 0.7
Required: Yes

Display Name: Max Tokens
Parameter: max_tokens
Component Type: Input
Default Value: 2048
Required: No

Advanced Parameters

4. Test Connection

4.1 Test Unsaved Configuration

After configuring the model, it's recommended to test the connection before saving.

Test Button

  • Located at drawer bottom, between "Cancel" and "Save" buttons
  • Button text: "Test Connection" / "Testing..."

Pre-test Validation

  • Check if provider and base model are selected
  • Check if required credential fields are filled
  • Shows warning if validation fails

Test Process

  1. Click "Test Connection" button
  2. Shows full-screen overlay: "Testing connectivity"
  3. Backend sends test request to verify configuration
  4. Test successful: Shows green success message
  5. Test failed: Shows red error message with specific reason

Test Significance

  • Verify if API Key is valid
  • Verify if Base URL is accessible
  • Verify if model name is correct
  • Avoid saving incorrect configuration

Test Connection

4.2 Test Saved Models

For saved models, you can quickly test connection status.

Trigger Method

  • Hover over model card
  • Click the cyan "link" icon

Test Result

  • Success: Shows green message "Connection OK", status tag updates to "Connected"
  • Failure: Shows red message "Connection Failed", status tag updates to "Unavailable"
  • List automatically refreshes after test completion

Best Practices

  • Regularly test model connectivity
  • Check configuration when "Unavailable" status is found
  • Re-test after changing API Key

Test Connection

5. Save Model Configuration

After successful connection test, click "Save" button to save configuration.

Pre-save Validation

  • Model name: Required, cannot be empty
  • Provider: Required, must select valid provider
  • Base model: Required, must select or enter model name
  • Credential fields: Check if all required fields are filled

Save Process

  1. Click "Save" button
  2. Button changes to "Saving..." with loading animation
  3. Shows full-screen overlay: "Processing, please wait..."
  4. Save successful: Shows green message "Saved"
  5. Close drawer, automatically refresh model list

Special Handling

  • When adding first model, automatically sets as user's default model
  • When editing the currently selected model, synchronously updates model info in memory

Save Configuration

6. Edit Model

6.1 Open Edit Interface

Hover over model card, click the blue "Edit" icon (pencil icon).

6.2 Modify Configuration

Pre-fill Logic

  • All fields pre-filled with current model's configuration values
  • Credential fields expanded
  • Advanced parameters loaded into table

Modifiable Items

  • Model name: Can rename
  • Base model: Can switch to other models from same provider
  • Credential fields: Can update API Key, Base URL, etc.
  • Advanced parameters: Can add, modify, delete parameters

Non-modifiable Items

  • Provider: Cannot change, delete and re-add if needed
  • Model situation (Public/Private): Automatically determined by provider

Edit Process

  1. Click edit icon
  2. Drawer slides out and loads model details
  3. Modify fields that need changes
  4. (Optional) Click "Test Connection" to verify changes
  5. Click "Save" button to save changes
  6. Close drawer, list auto-refreshes

Edit Model

7. Delete Model

7.1 Trigger Delete

Hover over model card, click the red "Delete" icon (trash icon).

7.2 Confirm Delete

Confirmation Dialog

  • Title: "Confirm Delete Model"
  • Content: "Are you sure you want to delete this model? This action cannot be undone."
  • Action buttons:
    • "Cancel" button: Gray, closes dialog
    • "Delete" button: Red danger button, executes delete

Delete Process

  1. Click "Delete" button in confirmation dialog
  2. Shows full-screen overlay: "Processing, please wait..."
  3. Delete successful shows green message: "Deleted"
  4. Auto-refresh model list
  5. If current page has no data after deletion, auto-jump to previous page

Notes

  • Delete operation cannot be undone, please operate carefully
  • Deleting model does not affect historical conversation records
  • If deleted model is currently in use, need to select another model

Delete Model

Delete Model

8. Supported Model Providers

CueMate supports multiple mainstream large language model providers. You can choose the appropriate model based on your needs:

8.1 Public Cloud Models

  1. OpenAI - World-leading AI company, provides GPT-5, GPT-4.1, GPT-4o and other advanced models, supports complex reasoning and multimodal understanding
  2. Anthropic - Claude series models, the latest Claude Sonnet 4.5 is the world's strongest programming model, excels at long text understanding and safe conversation
  3. Google Gemini - Google's latest multimodal model, supports text, image, and video understanding
  4. DeepSeek - Excellent Chinese model, high cost-effectiveness, strong coding ability, supports mathematical reasoning
  5. Moonshot (Kimi) - Moonshot AI, supports 200K ultra-long context, suitable for document analysis
  6. Zhipu AI - Tsinghua-affiliated, GLM-4 series models, excellent Chinese understanding
  7. Tongyi Qianwen - Alibaba Cloud Bailian platform, Qwen series models, comprehensive ecosystem
  8. Tencent Hunyuan - Tencent's self-developed LLM, strong Chinese ability, easy integration
  9. iFlytek Spark - iFlytek, optimized for voice scenarios, fluent Chinese conversation
  10. Volcano Engine - ByteDance, Doubao series models, multi-scenario adaptation
  11. SiliconFlow - Open source model inference platform, supports Llama, Qwen and other open source models
  12. Baidu Qianfan - Baidu's LLM platform, supports ERNIE-4.0, ERNIE-3.5 and other models
  13. Baichuan AI - Baichuan AI's Baichuan series models, supports Baichuan4, Baichuan3-Turbo, Baichuan3-Turbo-128k
  14. MiniMax - MiniMax's ultra-long text LLM, supports abab6.5-chat, abab6.5s-chat, up to 245K tokens context
  15. StepFun - Long context Step series models, supports step-1-8k, step-1-32k, step-1-256k, up to 256K tokens context
  16. SenseTime SenseNova - SenseTime's SenseNova series models, supports SenseChat-5, SenseChat-Turbo and 22 other models

8.2 Cloud Platform Model Services

  1. Azure OpenAI - Microsoft Azure cloud platform, provides enterprise-grade OpenAI model services, supports private network
  2. Amazon Bedrock - AWS managed service, integrates Claude, Llama and other models, secure and compliant
  3. Tencent Cloud - Tencent Cloud AI platform, provides Hunyuan and other model services

8.3 Private Deployment Models

  1. Ollama - Run open source models locally, supports Llama, Qwen, Gemma, data completely local
  2. vLLM - High-performance inference engine, supports large-scale model deployment, high throughput
  3. Xinference - Xorbits open source inference framework, supports various model formats, easy to extend
  4. Local Model - Custom local model service, runs completely offline, highest data security
  5. Regolo - Enterprise-grade private deployment solution, high availability guarantee, professional technical support

Selection Suggestions:

  • For Performance: OpenAI GPT-5, Claude Sonnet 4.5
  • For Programming: Claude Sonnet 4.5 (World's strongest), OpenAI GPT-5
  • High Cost-effectiveness: Claude Haiku 4.5, GPT-5 Mini, DeepSeek
  • Cost-conscious: DeepSeek, SiliconFlow, Ollama
  • Chinese Scenarios: Zhipu GLM-4, Tongyi Qianwen, Tencent Hunyuan
  • Data Security: Ollama, vLLM, Local Model
  • Enterprise Applications: Azure OpenAI, Amazon Bedrock, Private Deployment

9. Best Practices

Model Configuration Suggestions

Naming Convention

  • Use clear descriptive names, like "GPT-4 Turbo (Production)"
  • Distinguish same models for different purposes, like "Claude 3 Sonnet (Translation)", "Claude 3 Sonnet (Code)"
  • Avoid overly long names, recommend no more than 30 characters

Credential Security

  • API Key uses password mode after input
  • Don't expose API Key in screenshots or recordings
  • Regularly change API Key, re-test connection after change
  • Use different API Keys for different environments (dev/prod)

Parameter Optimization

  • Adjust temperature parameter based on use case:
    • Creative writing: 0.7-0.9 (more random)
    • Code generation: 0.2-0.4 (more deterministic)
    • Translation tasks: 0.3-0.5 (balanced)
  • Set max_tokens based on needs, avoid too large to prevent cost increase
  • Enable stream parameter for typewriter effect

Connectivity Management

  • Always test connection when adding new models
  • Regularly test connectivity of existing models
  • Investigate cause when "Unavailable" status is found
  • Keep backup models for main model failures

Efficiency Tips

Quick Add Model

  • Select specific provider node in left navigation tree
  • Click "Add Model" to directly enter configuration interface, skipping provider selection step

Search Filter

  • Use search function for quick location when you have many models
  • Combine left navigation and search box for precise filtering

Page Layout Optimization

  • Collapse left navigation when model count is small
  • Adjust items per page, recommend 12 or 18 for large screens

Common Troubleshooting

Connection Test Failed

  • Check if API Key is correct
  • Check if Base URL is accessible (for private deployment models)
  • Check if network connection is normal
  • Check if model name is correct

Save Failed

  • Confirm all required fields are filled
  • Check if model name is duplicate
  • Check if credential format is correct

Model Cannot Be Used

  • Confirm model status is "Connected"
  • Re-test connection
  • Check if API Key is expired
  • Check if balance is sufficient (public models)

Released under the GPL-3.0 License.