Skip to main content
MindStudio
Pricing
Blog About
My Workspace
Text Generation Model

Gemma 3.2

Google's most capable open multimodal model, combining image and text understanding with support for 140+ languages and a massive 128K context window.

Publisher Google
Type Text
Context Window 128,000 tokens
Training Data March 2025
Input $0.10/MTok
Output $0.20/MTok
Provider DeepInfra
LATESTREASONINGMULTI-MODAL

Open multimodal model with 128K context

Gemma 3 27B is an open-weight multimodal language model developed by Google DeepMind as the flagship model in the Gemma 3 family. It accepts both image and text inputs and generates text outputs, supporting over 140 languages and a context window of 128,000 tokens — sixteen times larger than the previous Gemma 2 generation. The model is built on the same research foundation as Google's Gemini models and was released in March 2025.

Gemma 3 27B is designed to run in resource-constrained environments, including on a single consumer GPU with 24GB of VRAM, as well as on laptops, desktops, and cloud infrastructure. It is well-suited for tasks such as visual question answering, document analysis, multilingual text generation, summarization, coding assistance, and logical reasoning. Its combination of multimodal input support, large context handling, and open-weight availability makes it a practical choice for developers building applications that require flexible deployment options.

What Gemma 3.2 supports

Multimodal Input

Processes both images and text as input in a single request, enabling tasks like visual question answering, image description, and document analysis.

128K Context Window

Handles up to 128,000 tokens of input per request, allowing analysis of long documents, large codebases, and extended multi-turn conversations.

Multilingual Generation

Generates and understands text in 140+ languages, making it suitable for globally-facing applications and cross-language tasks.

Reasoning & Analysis

Performs multi-step logical reasoning, summarization, and question answering across complex inputs including code and structured documents.

Code Generation

Generates, explains, and debugs code across common programming languages as part of its general text generation capabilities.

Flexible Deployment

Runs locally on a single GPU with 24GB VRAM (e.g., RTX 3090) as well as on cloud infrastructure, with open weights available for download.

Ready to build with Gemma 3.2?

Get Started Free

Benchmark scores

Scores represent accuracy — the percentage of questions answered correctly on each test.

Benchmark What it tests Score
MMLU-Pro Expert knowledge across 14 academic disciplines 66.9%
GPQA Diamond PhD-level science questions (biology, physics, chemistry) 42.8%
MATH-500 Undergraduate and competition-level math problems 88.3%
AIME 2024 American math olympiad problems 25.3%
LiveCodeBench Real-world coding tasks from recent competitions 13.7%
HLE Questions that challenge frontier models across many domains 4.7%
SciCode Scientific research coding and numerical methods 21.2%

Common questions about Gemma 3.2

What is the context window size for Gemma 3 27B?

Gemma 3 27B supports a context window of 128,000 tokens, which is sixteen times larger than the previous Gemma 2 generation.

Does Gemma 3 27B support image inputs?

Yes. Gemma 3 27B is a multimodal model that accepts both image and text as inputs, enabling tasks such as visual question answering, image description, and document analysis.

What languages does Gemma 3 27B support?

The model supports over 140 languages for both input understanding and text generation.

What is the training data cutoff for Gemma 3 27B?

Based on the available metadata, the model's training date is listed as March 2025. For precise knowledge cutoff details, refer to the official Gemma 3 Technical Report.

Can Gemma 3 27B be run locally?

Yes. The model is open-weight and can be deployed locally on a single GPU with 24GB of VRAM, such as an NVIDIA RTX 3090, as well as on laptops, desktops, or cloud infrastructure.

Is Gemma 3 27B free to use?

Gemma 3 27B is an open-weight model, meaning the weights are publicly available. Usage costs on MindStudio depend on the underlying inference provider (DeepInfra in this case); consult MindStudio's pricing page for current rates.

Parameters & options

Max Temperature 1
Max Response Size 8,000 tokens

Start building with Gemma 3.2

No API keys required. Create AI-powered workflows with Gemma 3.2 in minutes — free.