-
-
Notifications
You must be signed in to change notification settings - Fork 11.3k
Description
The model to consider.
Gemma 3 has a large, 128K context window, multilingual support in over 140 languages, and is available in more sizes than previous versions. Gemma 3 models are well-suited for a variety of text generation and image understanding tasks, including question answering, summarization, and reasoning.
Inputs and outputs
Input:
Text string, such as a question, a prompt, or a document to be summarized
Images, normalized to 896 x 896 resolution and encoded to 256 tokens each
Total input context of 128K tokens for the 4B, 12B, and 27B sizes, and 32K tokens for the 1B size
Output:
Generated text in response to the input, such as an answer to a question, analysis of image content, or a summary of a document
Total output context of 8192 tokens
https://huggingface.co/collections/google/gemma-3-release-67c6c6f89c4f76621268bb6d
https://storage.googleapis.com/deepmind-media/gemma/Gemma3Report.pdf
The closest model vllm already supports.
Gemma 2
What's your difficulty of supporting the model you want?
No response
Before submitting a new issue...
- Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the documentation page, which can answer lots of frequently asked questions.