Alpindale: Goliath 120B
alpindale/goliath-120b
Access Goliath 120B from Alpindale using Puter.js AI API.
Get Started// npm install @heyputer/puter.js
import { puter } from '@heyputer/puter.js';
puter.ai.chat("Explain quantum computing in simple terms", {
model: "alpindale/goliath-120b"
}).then(response => {
document.body.innerHTML = response.message.content;
});
<html>
<body>
<script src="https://js.puter.com/v2/"></script>
<script>
puter.ai.chat("Explain quantum computing in simple terms", {
model: "alpindale/goliath-120b"
}).then(response => {
document.body.innerHTML = response.message.content;
});
</script>
</body>
</html>
# pip install openai
from openai import OpenAI
client = OpenAI(
base_url="https://api.puter.com/puterai/openai/v1/",
api_key="YOUR_PUTER_AUTH_TOKEN",
)
response = client.chat.completions.create(
model="alpindale/goliath-120b",
messages=[
{"role": "user", "content": "Explain quantum computing in simple terms"}
],
)
print(response.choices[0].message.content)
curl https://api.puter.com/puterai/openai/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer YOUR_PUTER_AUTH_TOKEN" \
-d '{
"model": "alpindale/goliath-120b",
"messages": [
{"role": "user", "content": "Explain quantum computing in simple terms"}
]
}'
Model Card
Goliath 120B is a community-created large language model built by Alpindale by merging two fine-tuned Llama-2 70B models — Xwin and Euryale — into a single 120-billion-parameter model using the mergekit framework.
It was one of the earliest and most notable examples of the model-merging technique in the open-source LLM community, demonstrating that interleaving layers from two complementary fine-tunes could produce a capable larger model without traditional training. It supports Vicuna and Alpaca prompt formats, with Vicuna generally recommended.
Goliath 120B is primarily suited for creative writing, storytelling, and open-ended text generation. Its context window is limited to around 4–6K tokens, and no official benchmark scores have been published. Developers should consider it an experimental community model best fit for creative and conversational use cases rather than production workloads requiring verified performance.
Context Window 6K
tokens
Max Output 1K
tokens
Input Cost $3.75
per million tokens
Output Cost $7.5
per million tokens
Release Date Nov 5, 2023
Model Playground
Try Goliath 120B instantly in your browser.
This playground uses the Puter.js AI API — no API keys or setup required.
Frequently Asked Questions
You can access Goliath 120B by Alpindale through Puter.js AI API. Include the library in your web app or Node.js project and start making calls with just a few lines of JavaScript — no backend and no configuration required. You can also use it with Python or cURL via Puter's OpenAI-compatible API.
Yes, it is free if you're using it through Puter.js. With the User-Pays Model, you can add Goliath 120B to your app at no cost — your users pay for their own AI usage directly, making it completely free for you as a developer.
| Price per 1M tokens | |
|---|---|
| Input | $3.75 |
| Output | $7.5 |
Goliath 120B was created by Alpindale and released on Nov 5, 2023.
Goliath 120B supports a context window of 6K tokens. For reference, that is roughly equivalent to 12 pages of text.
Goliath 120B can generate up to 1K tokens in a single response.
Yes — the Goliath 120B API works with any JavaScript framework, Node.js, or plain HTML through Puter.js. Just include the library and start building. See the documentation for more details.
Get started with Puter.js
Add Goliath 120B to your app without worrying about API keys or setup.
Read the Docs View Tutorials