Skip to content
  1.  
  2. © 2023 – 2025 OpenRouter, Inc
    Favicon for google

    Google

    Browse models from Google

    37 models

    Tokens processed on OpenRouter

    • Google: Gemini 2.5 Flash Image (Nano Banana)Gemini 2.5 Flash Image (Nano Banana)
      3.42M tokens

      Gemini 2.5 Flash Image, a.k.a. "Nano Banana," is now generally available. It is a state of the art image generation model with contextual understanding. It is capable of image generation, edits, and multi-turn conversations. Aspect ratios can be controlled with the image_config API Parameter

      by google33K context$0.30/M input tokens$2.50/M output tokens
    $1.238/K input imgs
    $0.03/K output imgs
  3. Google: Gemini 2.5 Flash Preview 09-2025Gemini 2.5 Flash Preview 09-2025
    754M tokens

    Gemini 2.5 Flash Preview September 2025 Checkpoint is Google's state-of-the-art workhorse model, specifically designed for advanced reasoning, coding, mathematics, and scientific tasks. It includes built-in "thinking" capabilities, enabling it to provide responses with greater accuracy and nuanced context handling. Additionally, Gemini 2.5 Flash is configurable through the "max tokens for reasoning" parameter, as described in the documentation (https://openrouter.ai/docs/use-cases/reasoning-tokens#max-tokens-for-reasoning).

    by google1.05M context$0.30/M input tokens$2.50/M output tokens$1.238/K input imgs$1/M audio tokens
  4. Google: Gemini 2.5 Flash Lite Preview 09-2025Gemini 2.5 Flash Lite Preview 09-2025
    216M tokens

    Gemini 2.5 Flash-Lite is a lightweight reasoning model in the Gemini 2.5 family, optimized for ultra-low latency and cost efficiency. It offers improved throughput, faster token generation, and better performance across common benchmarks compared to earlier Flash models. By default, "thinking" (i.e. multi-pass reasoning) is disabled to prioritize speed, but developers can enable it via the Reasoning API parameter to selectively trade off cost for intelligence.

    by google1.05M context$0.10/M input tokens$0.40/M output tokens
  5. Google: Gemini 2.5 Flash Image Preview (Nano Banana)Gemini 2.5 Flash Image Preview (Nano Banana)
    19.7M tokens

    Gemini 2.5 Flash Image Preview, a.k.a. "Nano Banana," is a state of the art image generation model with contextual understanding. It is capable of image generation, edits, and multi-turn conversations.

    by google33K context$0.30/M input tokens$2.50/M output tokens$1.238/K input imgs$0.03/K output imgs
  6. Google: Gemini 2.5 Flash LiteGemini 2.5 Flash Lite
    4.47B tokens

    Gemini 2.5 Flash-Lite is a lightweight reasoning model in the Gemini 2.5 family, optimized for ultra-low latency and cost efficiency. It offers improved throughput, faster token generation, and better performance across common benchmarks compared to earlier Flash models. By default, "thinking" (i.e. multi-pass reasoning) is disabled to prioritize speed, but developers can enable it via the Reasoning API parameter to selectively trade off cost for intelligence.

    by google1.05M context$0.10/M input tokens$0.40/M output tokens
  7. Google: Gemma 3n 2BGemma 3n 2BFree variant
    7.94M tokens

    Gemma 3n E2B IT is a multimodal, instruction-tuned model developed by Google DeepMind, designed to operate efficiently at an effective parameter size of 2B while leveraging a 6B architecture. Based on the MatFormer architecture, it supports nested submodels and modular composition via the Mix-and-Match framework. Gemma 3n models are optimized for low-resource deployment, offering 32K context length and strong multilingual and reasoning performance across common benchmarks. This variant is trained on a diverse corpus including code, math, web, and multimodal data.

    by google8K context$0/M input tokens$0/M output tokens
  8. Google: Gemini 2.5 Flash Lite Preview 06-17Gemini 2.5 Flash Lite Preview 06-17
    4.56B tokens

    Gemini 2.5 Flash-Lite is a lightweight reasoning model in the Gemini 2.5 family, optimized for ultra-low latency and cost efficiency. It offers improved throughput, faster token generation, and better performance across common benchmarks compared to earlier Flash models. By default, "thinking" (i.e. multi-pass reasoning) is disabled to prioritize speed, but developers can enable it via the Reasoning API parameter to selectively trade off cost for intelligence.

    by google1.05M context$0.10/M input tokens$0.40/M output tokens$0.30/M audio tokens
  9. Google: Gemini 2.5 FlashGemini 2.5 Flash
    30.8B tokens

    Gemini 2.5 Flash is Google's state-of-the-art workhorse model, specifically designed for advanced reasoning, coding, mathematics, and scientific tasks. It includes built-in "thinking" capabilities, enabling it to provide responses with greater accuracy and nuanced context handling. Additionally, Gemini 2.5 Flash is configurable through the "max tokens for reasoning" parameter, as described in the documentation (https://openrouter.ai/docs/use-cases/reasoning-tokens#max-tokens-for-reasoning).

    by google1.05M context$0.30/M input tokens$2.50/M output tokens$1.238/K input imgs
  10. Google: Gemini 2.5 ProGemini 2.5 Pro
    19.2B tokens

    Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.

    by google1.05M context$1.25/M input tokens$10/M output tokens$5.16/K input imgs
  11. Google: Gemini 2.5 Pro Preview 06-05Gemini 2.5 Pro Preview 06-05

    Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.

    by google1.05M context$1.25/M input tokens$10/M output tokens$5.16/K input imgs
  12. Google: Gemma 1 2BGemma 1 2B

    Gemma 1 2B by Google is an open model built from the same research and technology used to create the Gemini models. Gemma models are well-suited for a variety of text generation tasks, including question answering, summarization, and reasoning. Usage of Gemma is subject to Google's Gemma Terms of Use.

    by google8K context
  13. Google: Gemma 3n 4BGemma 3n 4B
    105M tokens

    Gemma 3n E4B-it is optimized for efficient execution on mobile and low-resource devices, such as phones, laptops, and tablets. It supports multimodal inputs—including text, visual data, and audio—enabling diverse tasks such as text generation, speech recognition, translation, and image analysis. Leveraging innovations like Per-Layer Embedding (PLE) caching and the MatFormer architecture, Gemma 3n dynamically manages memory usage and computational load by selectively activating model parameters, significantly reducing runtime resource requirements. This model supports a wide linguistic range (trained in over 140 languages) and features a flexible 32K token context window. Gemma 3n can selectively load parameters, optimizing memory and computational efficiency based on the task or device capabilities, making it well-suited for privacy-focused, offline-capable applications and on-device AI solutions. Read more in the blog post

    by google32K context$0.02/M input tokens$0.04/M output tokens
  14. Google: Gemini 2.5 Flash Preview 05-20Gemini 2.5 Flash Preview 05-20

    Gemini 2.5 Flash May 20th Checkpoint is Google's state-of-the-art workhorse model, specifically designed for advanced reasoning, coding, mathematics, and scientific tasks. It includes built-in "thinking" capabilities, enabling it to provide responses with greater accuracy and nuanced context handling. Note: This model is available in two variants: thinking and non-thinking. The output pricing varies significantly depending on whether the thinking capability is active. If you select the standard variant (without the ":thinking" suffix), the model will explicitly avoid generating thinking tokens. To utilize the thinking capability and receive thinking tokens, you must choose the ":thinking" variant, which will then incur the higher thinking-output pricing. Additionally, Gemini 2.5 Flash is configurable through the "max tokens for reasoning" parameter, as described in the documentation (https://openrouter.ai/docs/use-cases/reasoning-tokens#max-tokens-for-reasoning).

    by google1.05M context
  15. Google: Gemini 2.5 Pro Preview 05-06Gemini 2.5 Pro Preview 05-06

    Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.

    by google1.05M context$1.25/M input tokens$10/M output tokens$5.16/K input imgs
  16. Google: Gemini 2.5 Flash Preview 04-17Gemini 2.5 Flash Preview 04-17

    Gemini 2.5 Flash is Google's state-of-the-art workhorse model, specifically designed for advanced reasoning, coding, mathematics, and scientific tasks. It includes built-in "thinking" capabilities, enabling it to provide responses with greater accuracy and nuanced context handling. Note: This model is available in two variants: thinking and non-thinking. The output pricing varies significantly depending on whether the thinking capability is active. If you select the standard variant (without the ":thinking" suffix), the model will explicitly avoid generating thinking tokens. To utilize the thinking capability and receive thinking tokens, you must choose the ":thinking" variant, which will then incur the higher thinking-output pricing. Additionally, Gemini 2.5 Flash is configurable through the "max tokens for reasoning" parameter, as described in the documentation (https://openrouter.ai/docs/use-cases/reasoning-tokens#max-tokens-for-reasoning).

    by google1.05M context
  17. Google: Gemini 2.5 Pro ExperimentalGemini 2.5 Pro Experimental

    This model has been deprecated by Google in favor of the (paid Preview model)[google/gemini-2.5-pro-preview]   Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.

    by google1.05M context
  18. Google: Gemma 3 1BGemma 3 1B

    Gemma 3 1B is the smallest of the new Gemma 3 family. It handles context windows up to 32k tokens, understands over 140 languages, and offers improved math, reasoning, and chat capabilities, including structured outputs and function calling. Note: Gemma 3 1B is not multimodal. For the smallest multimodal Gemma 3 model, please see Gemma 3 4B

    by google32K context
  19. Google: Gemma 3 4BGemma 3 4B
    482M tokens

    Gemma 3 introduces multimodality, supporting vision-language input and text outputs. It handles context windows up to 128k tokens, understands over 140 languages, and offers improved math, reasoning, and chat capabilities, including structured outputs and function calling.

    by google131K context$0.017/M input tokens$0.068/M output tokens
  20. Google: Gemma 3 12BGemma 3 12B
    49.9M tokens

    Gemma 3 introduces multimodality, supporting vision-language input and text outputs. It handles context windows up to 128k tokens, understands over 140 languages, and offers improved math, reasoning, and chat capabilities, including structured outputs and function calling. Gemma 3 12B is the second largest in the family of Gemma 3 models after Gemma 3 27B

    by google131K context$0.03/M input tokens$0.10/M output tokens
  21. Google: Gemma 3 27BGemma 3 27B
    873M tokens

    Gemma 3 introduces multimodality, supporting vision-language input and text outputs. It handles context windows up to 128k tokens, understands over 140 languages, and offers improved math, reasoning, and chat capabilities, including structured outputs and function calling. Gemma 3 27B is Google's latest open source model, successor to Gemma 2

    by google131K context$0.09/M input tokens$0.16/M output tokens$0.026/K input imgs
  22. Google: Gemini 2.0 Flash LiteGemini 2.0 Flash Lite
    3.21B tokens

    Gemini 2.0 Flash Lite offers a significantly faster time to first token (TTFT) compared to Gemini Flash 1.5, while maintaining quality on par with larger models like Gemini Pro 1.5, all at extremely economical token prices.

    by google1.05M context$0.075/M input tokens$0.30/M output tokens
  23. Google: Gemini 2.0 FlashGemini 2.0 Flash
    37B tokens

    Gemini Flash 2.0 offers a significantly faster time to first token (TTFT) compared to Gemini Flash 1.5, while maintaining quality on par with larger models like Gemini Pro 1.5. It introduces notable enhancements in multimodal understanding, coding capabilities, complex instruction following, and function calling. These advancements come together to deliver more seamless and robust agentic experiences.

    by google1M context$0.10/M input tokens$0.40/M output tokens$0.026/K input imgs$0.70/M audio tokens
  24. Google: Gemini 2.0 Flash ExperimentalGemini 2.0 Flash ExperimentalFree variant
    551M tokens

    Gemini Flash 2.0 offers a significantly faster time to first token (TTFT) compared to Gemini Flash 1.5, while maintaining quality on par with larger models like Gemini Pro 1.5. It introduces notable enhancements in multimodal understanding, coding capabilities, complex instruction following, and function calling. These advancements come together to deliver more seamless and robust agentic experiences.

    by google1.05M context$0/M input tokens$0/M output tokens
  25. Google: Gemini Experimental 1121Gemini Experimental 1121

    Experimental release (November 21st, 2024) of Gemini.

    by google41K context
  26. Google: Gemini Experimental 1114Gemini Experimental 1114

    Gemini 11-14 (2024) experimental model features "quality" improvements.

    by google41K context
  27. Google: Gemini 1.5 Flash 8BGemini 1.5 Flash 8B

    Gemini Flash 1.5 8B is optimized for speed and efficiency, offering enhanced performance in small prompt tasks like chat, transcription, and translation. With reduced latency, it is highly effective for real-time and large-scale operations. This model focuses on cost-effective solutions while maintaining high-quality results. Click here to learn more about this model. Usage of Gemini is subject to Google's Gemini Terms of Use.

    by google1M context
  28. Google: Gemini 1.5 Flash ExperimentalGemini 1.5 Flash Experimental

    Gemini 1.5 Flash Experimental is an experimental version of the Gemini 1.5 Flash model. Usage of Gemini is subject to Google's Gemini Terms of Use. #multimodal Note: This model is experimental and not suited for production use-cases. It may be removed or redirected to another model in the future.

    by google1M context
  29. Google: Gemini 1.5 Pro ExperimentalGemini 1.5 Pro Experimental

    Gemini 1.5 Pro Experimental is a bleeding-edge version of the Gemini 1.5 Pro model. Because it's currently experimental, it will be heavily rate-limited by Google. Usage of Gemini is subject to Google's Gemini Terms of Use. #multimodal

    by google1M context
  30. Google: Gemma 2 27BGemma 2 27B
    5.95M tokens

    Gemma 2 27B by Google is an open model built from the same research and technology used to create the Gemini models. Gemma models are well-suited for a variety of text generation tasks, including question answering, summarization, and reasoning. See the launch announcement for more details. Usage of Gemma is subject to Google's Gemma Terms of Use.

    by google8K context$0.65/M input tokens$0.65/M output tokens
  31. Google: Gemma 2 9BGemma 2 9B
    22.3M tokens

    Gemma 2 9B by Google is an advanced, open-source language model that sets a new standard for efficiency and performance in its size class. Designed for a wide variety of tasks, it empowers developers and researchers to build innovative applications, while maintaining accessibility, safety, and cost-effectiveness. See the launch announcement for more details. Usage of Gemma is subject to Google's Gemma Terms of Use.

    by google8K context$0.01/M input tokens$0.03/M output tokens
  32. Google: Gemini 1.5 Flash Gemini 1.5 Flash

    Gemini 1.5 Flash is a foundation model that performs well at a variety of multimodal tasks such as visual understanding, classification, summarization, and creating content from image, audio and video. It's adept at processing visual and text inputs such as photographs, documents, infographics, and screenshots. Gemini 1.5 Flash is designed for high-volume, high-frequency tasks where cost and latency matter. On most common tasks, Flash achieves comparable quality to other Gemini Pro models at a significantly reduced cost. Flash is well-suited for applications like chat assistants and on-demand content generation where speed and scale matter. Usage of Gemini is subject to Google's Gemini Terms of Use. #multimodal

    by google1M context
  33. Google: Gemini 1.5 ProGemini 1.5 Pro

    Google's latest multimodal model, supports image and video[0] in text or chat prompts. Optimized for language tasks including: - Code generation - Text generation - Text editing - Problem solving - Recommendations - Information extraction - Data extraction or generation - AI agents Usage of Gemini is subject to Google's Gemini Terms of Use. * [0]: Video input is not available through OpenRouter at this time.

    by google2M context
  34. Google: Gemma 7BGemma 7B

    Gemma by Google is an advanced, open-source language model family, leveraging the latest in decoder-only, text-to-text technology. It offers English language capabilities across text generation tasks like question answering, summarization, and reasoning. The Gemma 7B variant is comparable in performance to leading open source models. Usage of Gemma is subject to Google's Gemma Terms of Use.

    by google8K context
  35. Google: PaLM 2 Chat 32kPaLM 2 Chat 32k

    PaLM 2 is a language model by Google with improved multilingual, reasoning and coding capabilities.

    by google33K context
  36. Google: PaLM 2 Code Chat 32kPaLM 2 Code Chat 32k

    PaLM 2 fine-tuned for chatbot conversations that help with code-related questions.

    by google33K context
  37. Google: PaLM 2 ChatPaLM 2 Chat

    PaLM 2 is a language model by Google with improved multilingual, reasoning and coding capabilities.

    by google9K context
  38. Google: PaLM 2 Code ChatPaLM 2 Code Chat

    PaLM 2 fine-tuned for chatbot conversations that help with code-related questions.

    by google7K context