Description
GLM-4.5V is a vision-language foundation model for multimodal agent applications. Built on a Mixture-of-Experts (MoE) architecture with 106B parameters and 12B activated parameters, it achieves state-of-the-art results in video understanding, image Q&A, OCR, and document parsing, with strong gains in front-end web coding, grounding, and spatial reasoning. It offers a hybrid inference mode: a "thinking mode" for deep reasoning and a "non-thinking mode" for fast responses. Reasoning behavior can be toggled via the `reasoning` `enabled` boolean. [Learn more in our docs](https://openrouter.ai/docs/use-cases/reasoning-tokens#enable-reasoning-with-default-config)
API Usage Examples
OpenAI Compatible Endpoint
Use this endpoint with any OpenAI-compatible library. Model: Z.AI: GLM 4.5V (z-ai/glm-4.5v)
curl https://api.ridvay.com/v1/chat/completions -H "Content-Type: application/json" -H "Authorization: Bearer YOUR_API_KEY" -d '{
"model": "z-ai/glm-4.5v",
"messages": [
{
"role": "user",
"content": "Explain the capabilities of the Z.AI: GLM 4.5V model"
}
],
"temperature": 0.7,
"max_tokens": 1024
}'
Supported Modalities
- Text
- Images
API Pricing
- Input: 0.5$ / 1M tokens
- Output: 1.8$ / 1M tokens
Token Limits
- Max Output: 65,536 tokens
- Max Context: 65,536 tokens
Subscription Tiers
- free
- pro
- ultimate