⌘K

GLM-4.7

vtrix-glm-4.7

Open-source MoE model with 355B total/32B active parameters. 200K context window with 131K max output. Features three thinking modes (Interleaved, Preserved, Turn-level). Strong coding performance: 73.8% on SWE-bench, excellent multilingual coding support.

Authentication

authorization string required

All APIs require authentication via Bearer Token.

Get API Key:

Visit API Key Management Page to get your API Key.

Usage:

Add to request header:

Authorization: Bearer YOUR_API_KEY

Parameters

model string required

Model ID to use for the request.

Value: vtrix-glm-4.7


messages array required

Array of message objects


messages[].role string required

Message role

Options: user, assistant, system, developer


messages[].content string/array required

Text string or multimodal array


max_tokens integer

Maximum tokens to generate

Range: 1-131072


temperature number

Sampling temperature

Range: 0.0-2.0

Default: 1.0


top_p number

Nucleus sampling parameter

Range: 0.0-1.0

Default: 1.0


stream boolean

Stream response incrementally

Options: true, false

Default: false


Response Format

id string

Unique identifier for the completion


object string

Object type, always chat.completion


created integer

Unix timestamp of creation time


model string

The model used for completion


choices array

Array of completion choices

index integer

Choice index

message object

The generated message

role string

Message role (always assistant)

content string

Generated text content

finish_reason string

Reason for completion termination


usage object

Token usage statistics

prompt_tokens integer

Number of tokens in the prompt

completion_tokens integer

Number of tokens in the completion

total_tokens integer

Total tokens used