Google Gemini 2.0 Flash Lite
Google Gemini 2.0 Flash Lite
Google Gemini 2.0 Flash Lite is a multimodal model designed to handle lightweight tasks. It is designed for high-volume, low-latency tasks, making it highly efficient for large-scale use cases like summarization, multimodal processing, and categorization but with higher quality than Gemini 1.5 Flash.
Model details
Item | Value | Description |
---|---|---|
Model name | Google Gemini 2.0 Flash Lite | The name of the model. |
API model name | google__gemini_2_0_flash_lite_preview | The name of the model that is used in the Box AI API for model overrides. The user must provide this exact name for the API to work. |
Hosting layer | The trusted organization that securely hosts LLM. | |
Model provider | The organization that provides this model. | |
Release date | February 5th 2025 | The release date for the model. |
Knowledge cutoff date | June 2024 | The date after which the model does not get any information updates. |
Input context window | 1m tokens | The number of tokens supported by the input context window. |
Maximum output tokens | 8k tokens | The number of tokens that can be generated by the model in a single request. |
Empirical throughput | 168 | The number of tokens the model can generate per second. |
Open source | No | Specifies if the model's code is available for public use. |
Additional documentation
For additional information, see official Google Gemini 2.0 Flash Lite documentation.