Blog

GLM 4.7 Flash Is Now Available in Puter.js

On this page

Puter.js now supports GLM-4.7-Flash, Z.AI's lightweight and high-speed variant of their flagship GLM-4.7 model.

What is GLM 4.7 Flash?

GLM 4.7 Flash is designed for speed and efficiency while maintaining strong performance. It features a 200K token context window, making it suitable for processing long documents and generating extended responses.

The model achieves open-source state-of-the-art scores among models of comparable size on major benchmarks including SWE-bench Verified. It excels at both frontend and backend programming tasks, as well as general-purpose applications like writing, translation, and long-form text processing.

Examples

Basic Chat

puter.ai.chat("Explain the difference between REST and GraphQL", {
    model: "z-ai/glm-4.7-flash"
})

Code Generation

puter.ai.chat(`Write a Python function that implements binary search
with proper error handling and type hints`,
    { model: "z-ai/glm-4.7-flash" }
)

Long Context Processing

puter.ai.chat(`Summarize the key points from this technical document:
${longDocument}`,
    { model: "z-ai/glm-4.7-flash" }
)

Get Started Now

Just add one script tag to your HTML:

<script src="https://js.puter.com/v2/"></script>

No API keys or account needed. Start building with GLM 4.7 Flash immediately.

Learn more:

Free, Serverless AI and Cloud

Start creating powerful web applications with Puter.js in seconds!

Get Started Now

Read the Docs Try the Playground