Deep Cogito

Deep Cogito API

Access Deep Cogito instantly with Puter.js, and add AI to any app in a few lines of code without backend or API keys.

// npm install @heyputer/puter.js
import { puter } from '@heyputer/puter.js';

puter.ai.chat("Explain AI like I'm five!", {
    model: "deepcogito/cogito-v2-preview-llama-70b"
}).then(response => {
    console.log(response);
});
<html>
<body>
    <script src="https://js.puter.com/v2/"></script>
    <script>
        puter.ai.chat("Explain AI like I'm five!", {
            model: "deepcogito/cogito-v2-preview-llama-70b"
        }).then(response => {
            console.log(response);
        });
    </script>
</body>
</html>

List of Deep Cogito Models

Chat

Deep Cogito: Cogito v2.1 671B

deepcogito/cogito-v2.1-671b

Cogito v2.1 671B is DeepCogito's flagship mixture-of-experts model with 671B total parameters (37B activated), ranking among the most powerful open-source AI models globally. It matches or exceeds DeepSeek R1 performance while using 60% shorter reasoning chains, supports 128k context and 30+ languages.

Chat

Cogito V2 Preview Llama 109B

deepcogito/cogito-v2-preview-llama-109b-moe

Cogito v2 109B MoE is a sparse mixture-of-experts model built on Llama architecture that uses selective expert activation for efficient complex reasoning tasks. It features hybrid reasoning with both standard and self-reflective modes, achieving strong performance while being cost-effective at $0.18/$0.59 per million input/output tokens.

Chat

Deep Cogito: Cogito V2 Preview Llama 405B

deepcogito/cogito-v2-preview-llama-405b

Cogito v2 405B is a dense large-scale model built on Llama architecture, optimized for high performance across varied hardware setups. It offers hybrid reasoning capabilities and is designed to approach frontier-level intelligence within its size class through improved non-reasoning mode intuition.

Chat

Deep Cogito: Cogito V2 Preview Llama 70B

deepcogito/cogito-v2-preview-llama-70b

Cogito v2 70B is a dense model built on Llama architecture, ideal for low-latency applications and fine-tuning scenarios. It features dual-mode operation (standard LLM or self-reflective reasoning) and is optimized for environments with limited GPU capacity at $0.88 per million tokens.

Frequently Asked Questions

What is this Deep Cogito API about?

The Deep Cogito API gives you access to models for AI chat. Through Puter.js, you can start using Deep Cogito models instantly with zero setup or configuration.

Which Deep Cogito models can I use?

Puter.js supports a variety of Deep Cogito models, including Deep Cogito: Cogito v2.1 671B, Cogito V2 Preview Llama 109B, Deep Cogito: Cogito V2 Preview Llama 405B, and more. Find all AI models supported by Puter.js in the AI model list.

How much does it cost?

With the User-Pays model, users cover their own AI costs through their Puter account. This means you can build apps without worrying about infrastructure expenses.

What is Puter.js?

Puter.js is a JavaScript library that provides access to AI, storage, and other cloud services directly from a single API. It handles authentication, infrastructure, and scaling so you can focus on building your app.

Does this work with React / Vue / Vanilla JS / Node / etc.?

Yes — the Deep Cogito API through Puter.js works with any JavaScript framework, Node.js, or plain HTML. Just include the library and start building. See the documentation for more details.