Aion Labs: Aion-1.0-Mini

Access Aion-1.0-Mini from Aion Labs using Puter.js AI API.

Get Started

Model Card

Aion 1.0 Mini is a 32B-parameter reasoning model from AionLabs, distilled from DeepSeek-R1 and based on a modified FuseAI variant. It is designed for strong performance in mathematics, coding, and logic at a fraction of the cost of full-scale models.

It delivers standout speed and pricing, consistently ranking among the fastest and most affordable options available. On Benchable, it scored 99.0% on email classification and 82.0% on reasoning tasks.

This model is a good fit for developers who need fast, budget-friendly reasoning for structured tasks and can work around its instruction-following limitations.

Context Window 131K

tokens

Max Output 33K

tokens

Input Cost $0.7

per million tokens

Output Cost $1.4

per million tokens

Release Date Feb 1, 2025

 

API Usage Example

Add Aion-1.0-Mini to your app with just a few lines of code.
No backend, no configuration required.

// npm install @heyputer/puter.js
import { puter } from '@heyputer/puter.js';

puter.ai.chat("Explain quantum computing in simple terms", {
    model: "aion-labs/aion-1.0-mini"
}).then(response => {
    document.body.innerHTML = response.message.content;
});
<html>
<body>
    <script src="https://js.puter.com/v2/"></script>
    <script>
        puter.ai.chat("Explain quantum computing in simple terms", {
            model: "aion-labs/aion-1.0-mini"
        }).then(response => {
            document.body.innerHTML = response.message.content;
        });
    </script>
</body>
</html>
# pip install openai
from openai import OpenAI

client = OpenAI(
    base_url="https://api.puter.com/puterai/openai/v1/",
    api_key="YOUR_PUTER_AUTH_TOKEN",
)

response = client.chat.completions.create(
    model="aion-labs/aion-1.0-mini",
    messages=[
        {"role": "user", "content": "Explain quantum computing in simple terms"}
    ],
)

print(response.choices[0].message.content)
curl https://api.puter.com/puterai/openai/v1/chat/completions \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer YOUR_PUTER_AUTH_TOKEN" \
  -d '{
    "model": "aion-labs/aion-1.0-mini",
    "messages": [
      {"role": "user", "content": "Explain quantum computing in simple terms"}
    ]
  }'

View full documentation →

More AI Models From Aion Labs

Chat

Aion-2.0

Aion 2.0 is a fine-tuned variant of DeepSeek V3.2, developed by AionLabs and optimized for immersive roleplaying and storytelling. It excels at generating narratives with natural tension, conflict, and dramatic stakes, and handles mature or darker themes with notable nuance. The model offers a 131K-token context window with up to 32K tokens of output, making it well-suited for long-form creative sessions. It supports function calling and streaming. On third-party benchmarks, it has scored 99.5% on general knowledge, 96% on mathematics, and 93.5% on coding tasks. Aion 2.0 is a strong pick for developers building interactive fiction, character-driven chat experiences, or creative writing tools where narrative depth and engagement matter more than raw speed.

Chat

Aion-1.0

Aion 1.0 is AionLabs' most powerful reasoning model, a multi-model system built on DeepSeek-R1 and augmented with Tree of Thoughts (ToT) and Mixture of Experts (MoE) techniques. It supports a 131K context window with up to 32K output tokens and includes vision capabilities. The model excels at reasoning and coding tasks, scoring 96.0% on coding, 99.5% on general knowledge, and achieving perfect accuracy on reasoning and email classification benchmarks (Benchable). It also ranks among the fastest models at its price point. Best suited for developers who need strong reasoning, coding assistance, and classification at competitive throughput.

Chat

Aion-RP 1.0 (8B)

Aion RP 1.0 8B is an uncensored roleplay and creative writing model from AionLabs, fine-tuned from the Llama 3.1 8B base model rather than an instruct variant. This base-model approach is designed to produce more natural and varied writing. It ranks highest in the character evaluation portion of RPBench-Auto, a roleplaying-specific benchmark derived from Arena-Hard-Auto where LLMs evaluate each other's responses. The model supports the full 131K context window and multi-turn conversations. Best suited for character-driven chat applications, interactive storytelling, and persona-consistent dialogue. A recommended temperature of 0.7 is advised, as higher values can degrade output quality.

View all Aion Labs models →

Frequently Asked Questions

How do I use Aion-1.0-Mini?

You can access Aion-1.0-Mini by Aion Labs through Puter.js AI API. Include the library in your web app or Node.js project and start making calls with just a few lines of JavaScript — no backend and no configuration required. You can also use it with Python or cURL via Puter's OpenAI-compatible API.

Is Aion-1.0-Mini free?

Yes, it is free if you're using it through Puter.js. With the User-Pays Model, you can add Aion-1.0-Mini to your app at no cost — your users pay for their own AI usage directly, making it completely free for you as a developer.

What is the pricing for Aion-1.0-Mini?
Pricing for Aion-1.0-Mini is based on the number of input and output tokens used per request.
Price per 1M tokens
Input$0.7
Output$1.4
Who created Aion-1.0-Mini?

Aion-1.0-Mini was created by Aion Labs and released on Feb 1, 2025.

What is the context window of Aion-1.0-Mini?

Aion-1.0-Mini supports a context window of 131K tokens. For reference, that is roughly equivalent to 262 pages of text.

What is the max output length of Aion-1.0-Mini?

Aion-1.0-Mini can generate up to 33K tokens in a single response.

Does it work with React / Vue / Vanilla JS / Node / etc.?

Yes — the Aion-1.0-Mini API works with any JavaScript framework, Node.js, or plain HTML through Puter.js. Just include the library and start building. See the documentation for more details.

Get started with Puter.js

Add Aion-1.0-Mini to your app without worrying about API keys or setup.

Read the Docs View Tutorials