You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
**Google Gemini 2.0 Flash** is a multimodal model designed for optimal for high-volume, high-frequency tasks at scale. It capable of multimodal reasoning and has a context window of 1 million tokens.
14
+
15
+
## Model details
16
+
17
+
| Item | Value | Description |
18
+
|-----------|----------|----------|
19
+
|Model name|**Google Gemini 2.0 Flash**| The name of the model. |
20
+
|API model name|`google__gemini_2_0_flash_001`| The name of the model that is used in the [Box AI API for model overrides][overrides]. The user must provide this exact name for the API to work. |
21
+
|Hosting layer|**Google**| The trusted organization that securely hosts LLM. |
22
+
|Model provider|**Google**| The organization that provides this model. |
23
+
|Release date|**February 5th 2025**| The release date for the model.|
24
+
|Knowledge cutoff date|**June 2024**| The date after which the model does not get any information updates. |
25
+
|Input context window |**1m tokens**| The number of tokens supported by the input context window.|
26
+
|Maximum output tokens |**8k tokens**|The number of tokens that can be generated by the model in a single request.|
27
+
|Empirical throughput|**168**| The number of tokens the model can generate per second.|
28
+
|Open source |**No**| Specifies if the model's code is available for public use.|
29
+
30
+
## Additional documentation
31
+
32
+
For additional information, see [official Google Gemini 2.0 Flash documentation][vertex-ai-gemini-models].
**Google Gemini 2.0 Flash Lite** is a multimodal model designed to handle lightweight tasks. It is designed for high-volume, low-latency tasks, making it highly efficient for large-scale use cases like summarization, multimodal processing, and categorization but with higher quality than Gemini 1.5 Flash.
14
+
15
+
## Model details
16
+
17
+
| Item | Value | Description |
18
+
|-----------|----------|----------|
19
+
|Model name|**Google Gemini 2.0 Flash Lite**| The name of the model. |
20
+
|API model name|`google__gemini_2_0_flash_lite_preview`| The name of the model that is used in the [Box AI API for model overrides][overrides]. The user must provide this exact name for the API to work. |
21
+
|Hosting layer|**Google**| The trusted organization that securely hosts LLM. |
22
+
|Model provider|**Google**| The organization that provides this model. |
23
+
|Release date|**February 5th 2025**| The release date for the model.|
24
+
|Knowledge cutoff date|**June 2024**| The date after which the model does not get any information updates. |
25
+
|Input context window |**1m tokens**| The number of tokens supported by the input context window.|
26
+
|Maximum output tokens |**8k tokens**|The number of tokens that can be generated by the model in a single request.|
27
+
|Empirical throughput|**168**| The number of tokens the model can generate per second.|
28
+
|Open source |**No**| Specifies if the model's code is available for public use.|
29
+
30
+
## Additional documentation
31
+
32
+
For additional information, see [official Google Gemini 2.0 Flash Lite documentation][vertex-ai-gemini-models].
0 commit comments