Google Gemini 2.0 Flash Lite
Google Gemini 2.0 Flash Lite
Google Gemini 2.0 Flash Lite is a multimodal model designed to handle lightweight tasks. It is designed for high-volume, low-latency tasks, making it highly efficient for large-scale use cases like summarization, multimodal processing, and categorization but with higher quality than Gemini 1.5 Flash.
Model details
| Item | Value | Description |
|---|---|---|
| Model name | Google Gemini 2.0 Flash Lite | The name of the model. |
| Model category | Standard | The category of the model - standard or premium. |
| API model name | google__gemini_2_0_flash_lite_preview | The name of the model that is used in the Box AI API for model overrides. The user must provide this exact name for the API to work. |
| Hosting layer | The trusted organization that securely hosts LLM. | |
| Model provider | The organization that provides this model. | |
| Release date | February 5th 2025 | The release date for the model. |
| Knowledge cutoff date | June 2024 | The date after which the model does not get any information updates. |
| Input context window | 1m tokens | The number of tokens supported by the input context window. |
| Maximum output tokens | 8k tokens | The number of tokens that can be generated by the model in a single request. |
| Empirical throughput | 168 | The number of tokens the model can generate per second. |
| Open source | No | Specifies if the model's code is available for public use. |
Additional documentation
For additional information, see official Google Gemini 2.0 Flash Lite documentation.