Inception API
Access Inception instantly with Puter.js, and add AI to any app in a few lines of code without backend or API keys.
// npm install @heyputer/puter.js
import { puter } from '@heyputer/puter.js';
puter.ai.chat("Explain AI like I'm five!", {
model: "inception/mercury-2"
}).then(response => {
console.log(response);
});
<html>
<body>
<script src="https://js.puter.com/v2/"></script>
<script>
puter.ai.chat("Explain AI like I'm five!", {
model: "inception/mercury-2"
}).then(response => {
console.log(response);
});
</script>
</body>
</html>
List of Inception Models
Mercury 2
inception/mercury-2
Mercury 2 is a diffusion-based reasoning language model from Inception Labs that refines all tokens in parallel rather than generating them sequentially, achieving over 1,000 tokens per second — roughly 5x faster than speed-optimized competitors like Claude Haiku and GPT-5 Mini at comparable quality. On reasoning benchmarks, Mercury 2 scores 91.1 on AIME 2025 and 73.6 on GPQA. It also placed second on the Copilot Arena leaderboard for quality while ranking first for speed overall. With a 128K context window, it is purpose-built for latency-sensitive applications — real-time assistants, high-throughput pipelines, and cost-conscious production workloads where reasoning capability matters.
ChatMercury Coder
inception/mercury-coder
Mercury Coder is a code-specialized diffusion language model from Inception Labs, built on the same parallel token refinement architecture as Mercury. It is available in Mini and Small sizes. On fill-in-the-middle tasks, Mercury Coder Small scored 84.8% average accuracy, exceeding Codestral 2501 (82.5%). On MultiPL-E, it reaches 82.0% in C++, 83.9% in JavaScript, and 82.6% in TypeScript. In Copilot Arena human evaluations, Mercury Coder Mini ranked second in user preference with an average latency of just 25 milliseconds. It is the go-to choice for real-time code completion, autocomplete, and apply-edit workflows where both speed and accuracy are critical.
ChatMercury
inception/mercury
Mercury is the world's first commercial-scale diffusion large language model from Inception Labs. It generates text through iterative parallel refinement rather than sequential token prediction, enabling dramatically higher throughput without sacrificing output quality. It matches the performance of frontier speed-optimized models such as GPT-4o Mini and Gemini 1.5 Flash across knowledge, coding, instruction-following, and math benchmarks, while running up to 10x faster. It is OpenAI API-compatible for straightforward integration. Mercury is well-suited for API use cases that demand high concurrency, fast response times, or cost efficiency — including chat, summarization, and general-purpose text generation at scale.
Frequently Asked Questions
The Inception API gives you access to models for AI chat. Through Puter.js, you can start using Inception models instantly with zero setup or configuration.
Puter.js supports a variety of Inception models, including Mercury 2, Mercury Coder, and Mercury. Find all AI models supported by Puter.js in the AI model list.
With the User-Pays model, users cover their own AI costs through their Puter account. This means you can build apps without worrying about infrastructure expenses.
Puter.js is a JavaScript library that provides access to AI, storage, and other cloud services directly from a single API. It handles authentication, infrastructure, and scaling so you can focus on building your app.
Yes — the Inception API through Puter.js works with any JavaScript framework, Node.js, or plain HTML. Just include the library and start building. See the documentation for more details.