Blog

Qwen 3.5 Models Are Now Available in Puter.js

On this page

Puter.js now supports the full Qwen 3.5 model series from Alibaba's Qwen team—six models spanning from a 3B-active ultra-efficient Flash model to a 397B-parameter flagship that competes with GPT-5.2 and Claude Opus 4.5.

What is Qwen 3.5?

Qwen 3.5 is the latest model family from Alibaba's Qwen team. All models use a hybrid Gated DeltaNet + MoE architecture with native multimodal support, up to 1M token context, and coverage of 201 languages.

  • Qwen3.5-397B-A17B: The open-weight flagship with 397B total parameters and 17B active per token. Competes with GPT-5.2 and Claude Opus 4.5 on reasoning, coding, and multimodal benchmarks.
  • Qwen3.5-Plus: The hosted flagship API with a 1M token context window and built-in tool use including code interpreter. Designed for agentic workflows.
  • Qwen3.5-122B-A10B: The largest medium MoE model (122B total, 10B active). Leads the lineup on tool-use and agent benchmarks like BFCL-V4 and BrowseComp.
  • Qwen3.5-27B: The only dense model in the series. All 27B parameters active on every pass. Ties GPT-5 mini on SWE-bench Verified at 72.4 and runs well on consumer hardware.
  • Qwen3.5-35B-A3B: A sparse MoE model activating just 3B of 35B parameters. Outperforms the previous-gen 235B flagship and runs on GPUs with as little as 8 GB VRAM.
  • Qwen3.5-Flash: The production API version of 35B-A3B with 1M context and native function calling at ~$0.10/M input tokens.

Examples

General reasoning

puter.ai.chat("Explain the trade-offs between event sourcing and traditional CRUD for a banking application",
  { model: 'qwen/qwen3.5-397b-a17b' }
);

Long-context analysis

puter.ai.chat("Analyze this codebase and identify potential memory leaks:\n" + largeCodebase,
  { model: 'qwen/qwen3.5-plus-02-15' }
);

Code generation

puter.ai.chat("Write a Rust implementation of a lock-free concurrent hash map",
  { model: 'qwen/qwen3.5-27b' }
);

Streaming

const response = await puter.ai.chat(
  "Write a step-by-step guide to deploying a Next.js app on AWS Lambda",
  { model: 'qwen/qwen3.5-flash-02-23', stream: true }
);

for await (const part of response) {
  puter.print(part?.text);
}

Get Started Now

// npm install @heyputer/puter.js
import { puter } from '@heyputer/puter.js';

Or add one script tag to your HTML:

<script src="https://js.puter.com/v2/"></script>

No API keys needed. Start building with Qwen 3.5 models immediately.

Learn more:

Free, Serverless AI and Cloud

Start creating powerful web applications with Puter.js in seconds!

Get Started Now

Read the Docs Try the Playground