Google Releases Gemini 2.5 Flash-Lite AI Model

Google has now made its Gemini 2.5 Flash-Lite AI model publicly available to developers. This version is the fastest and most cost-effective model in the Gemini 2.5 series. After completing a one-month preview period, it is now accessible via AI Studio and Vertex AI.

According to information published on Google’s developer blog, the usage cost for this model is $0.10 per million input tokens and $0.40 per million output tokens. Despite its lightweight nature, the model retains advanced capabilities in coding, mathematics, science, reasoning, and multimodal understanding, delivering impressive overall performance.

Gemini 2.5 Flash-Lite Features for Google Customers

To showcase its model’s capabilities, Google tested Gemini 2.5 Flash-Lite in various real-world scenarios. One such scenario involved the company Satlyt, which uses Gemini to speed up satellite data processing and summarize telemetry information. According to Google, the model has also helped Satlyt reduce its energy consumption by around 30%.

Other companies, such as HeyGen, are using the model to translate their video content into over 180 languages.

The Gemini 2.5 Flash-Lite AI model was first introduced in June during the public rollout of the Gemini 2.5 Flash and 2.5 Pro models. From the beginning, Google positioned this model as a fast and affordable option for developers, whereas the 2.5 Pro model is designed for heavier workloads.