Gemini 2.0 Flash-Lite API

Access Gemini 2.0 Flash-Lite from Google using Puter.js AI API.

Get Started

Model Card

Gemini 2.0 Flash-Lite is Google's most cost-efficient model, optimized for large-scale text output tasks. It offers simplified pricing and lower costs than Flash while maintaining solid performance for high-volume workloads.

Context Window 1M

tokens

Max Output 8K

tokens

Input Cost $0.08

per million tokens

Output Cost $0.3

per million tokens

Input text, image, audio, video, pdf

modalities

Tool Use Yes

 

Knowledge Cutoff Jun 2024

 

Release Date Dec 11, 2024

 

API Usage Example

Add Gemini 2.0 Flash-Lite to your app with just a few lines of code.
No API keys, no backend, no configuration required.

// npm install @heyputer/puter.js
import { puter } from '@heyputer/puter.js';

puter.ai.chat("Explain quantum computing in simple terms", {
    model: "google/gemini-2.0-flash-lite"
}).then(response => {
    document.body.innerHTML = response.message.content;
});
<html>
<body>
    <script src="https://js.puter.com/v2/"></script>
    <script>
        puter.ai.chat("Explain quantum computing in simple terms", {
            model: "google/gemini-2.0-flash-lite"
        }).then(response => {
            document.body.innerHTML = response.message.content;
        });
    </script>
</body>
</html>

View full documentation →

Frequently Asked Questions

What is this Gemini 2.0 Flash-Lite API about?

The Gemini 2.0 Flash-Lite API gives you access to Google's chat model through Puter.js. With just a few lines of JavaScript, you can integrate Gemini 2.0 Flash-Lite into any web app or Node.js project — no API keys, no backend, and no configuration required.

Who created Gemini 2.0 Flash-Lite?

Gemini 2.0 Flash-Lite was created by Google and released on Dec 11, 2024.

What is the context window of Gemini 2.0 Flash-Lite?

Gemini 2.0 Flash-Lite supports a context window of 1M tokens. For reference, that is roughly equivalent to 2,097 pages of text.

What is the max output length of Gemini 2.0 Flash-Lite?

Gemini 2.0 Flash-Lite can generate up to 8K tokens in a single response.

What is the knowledge cutoff of Gemini 2.0 Flash-Lite?

Gemini 2.0 Flash-Lite has a knowledge cutoff date of Jun 2024. This means the model was trained on data available up to that date.

What types of input can Gemini 2.0 Flash-Lite process?

Gemini 2.0 Flash-Lite accepts the following input types: text, image, audio, video, pdf. It produces: text.

Does Gemini 2.0 Flash-Lite support tool use (function calling)?

Yes, Gemini 2.0 Flash-Lite supports tool use (function calling), allowing it to interact with external tools, APIs, and data sources as part of its response flow.

How much does it cost?
The Gemini 2.0 Flash-Lite API is available through the User-Pays Model. As a developer, you can add the Gemini 2.0 Flash-Lite API to your app for free — your users pay for their own AI costs directly.
Price per 1M tokens
Input$0.08
Output$0.3
How do I access the Gemini 2.0 Flash-Lite API?

You can access the Gemini 2.0 Flash-Lite API with just a few lines of JavaScript — no API keys, no backend, and no configuration required. Include the Puter.js library in your project and start making calls right away. For more details, check out our documentation.

Does the Gemini 2.0 Flash-Lite API work with React / Vue / Vanilla JS / Node / etc.?

Yes — the Gemini 2.0 Flash-Lite API works with any JavaScript framework, Node.js, or plain HTML through Puter.js. Just include the library and start building. See the documentation for more details.

Get started with Puter.js

Add Gemini 2.0 Flash-Lite to your app without worrying about API keys or setup.

Read the Docs View Tutorials