Cogito V2 Preview Llama 109B API
Access Cogito V2 Preview Llama 109B from Deep Cogito using Puter.js AI API.
Get Starteddeepcogito/cogito-v2-preview-llama-109b-moe
Model Card
Cogito v2 109B MoE is a sparse mixture-of-experts model built on Llama architecture that uses selective expert activation for efficient complex reasoning tasks. It features hybrid reasoning with both standard and self-reflective modes, achieving strong performance while being cost-effective at $0.18/$0.59 per million input/output tokens.
Context Window
N/A
tokens
Max Output
N/A
tokens
Input Cost
$0.18
per million tokens
Output Cost
$0.59
per million tokens
API Usage Example
Add Cogito V2 Preview Llama 109B to your app with just a few lines of code.
No API keys, no backend, no configuration required.
<html>
<body>
<script src="https://js.puter.com/v2/"></script>
<script>
puter.ai.chat("Explain quantum computing in simple terms", {
model: "deepcogito/cogito-v2-preview-llama-109b-moe"
}).then(response => {
document.body.innerHTML = response.message.content;
});
</script>
</body>
</html>
Get started with Puter.js
Add Cogito V2 Preview Llama 109B to your app without worrying about API keys or setup.
Read the Docs View Tutorials