Search

335 results

Clear filters
  • JULY 16, 2025 / AI

    Unlock Gemini’s reasoning: A step-by-step guide to logprobs on Vertex AI

    The `logprobs` feature has been officially introduced in the Gemini API on Vertex AI, provides insight into the model's decision-making by showing probability scores for chosen and alternative tokens. This step-by-step guide will walk you through how to enable and interpret this feature and apply it to powerful use cases such as confident classification, dynamic autocomplete, and quantitative RAG evaluation.

    logprobs_meta
  • JULY 16, 2025 / Cloud

    Stanford’s Marin foundation model: The first fully open model developed using JAX

    The Marin project aims to expand the definition of 'open' in AI to include the entire scientific process, not just the model itself, by making the complete development journey accessible and reproducible. This effort, powered by the JAX framework and its Levanter tool, allows for deep scrutiny, trust in, and building upon foundation models, fostering a more transparent future for AI research.

    Stanford Marin project in JAX
  • JULY 16, 2025 / Gemini

    Simplify your Agent "vibe building" flow with ADK and Gemini CLI

    The updated Agent Development Kit (ADK) simplifies and accelerates the process of building AI agents by providing the CLI with a deep, cost-effective understanding of the ADK framework, allowing developers to quickly ideate, generate, test, and improve functional agents through conversational prompts, eliminating friction and keeping them in a productive "flow" state.

    ADK + Gemini CLI: Supercharge Your Agent Building Vibe
  • JULY 14, 2025 / Cloud

    Enterprise truth in action: Apigee API hub fueling powerful Developer Portals

    The Apigee API hub and Developer Portals are distinct but interconnected parts of the Apigee platform that help organizations discover and manage APIs for different personas, unlocking API potential and accelerating innovation.

    Enterprise Truth in Action: API hub Fueling Powerful Developer Portals
  • JULY 14, 2025 / Gemini

    Gemini Embedding now generally available in the Gemini API

    The Gemini Embedding text model is now generally available in the Gemini API and Vertex AI. This versatile model has consistently ranked #1 on the MTEB Multilingual leaderboard since its experimental launch in March, supports over 100 languages, has a 2048 maximum input token length, and is priced at $0.15 per 1M input tokens.

    Gemini Embedding now generally available in the Gemini API
  • JULY 10, 2025 / Gemini

    Announcing GenAI Processors: Build powerful and flexible Gemini applications

    GenAI Processors is a new open-source Python library from Google DeepMind designed to simplify the development of AI applications, especially those handling multimodal input and requiring real-time responsiveness, by providing a consistent "Processor" interface for all steps from input handling to model calls and output processing, for seamless chaining and concurrent execution.

    Announcing GenAI Processors: Streamline your Gemini application development
  • JULY 10, 2025 / Cloud

    Advancing agentic AI development with Firebase Studio

    Updates in Firebase Studio include new Agent modes, foundational support for the Model Context Protocol (MCP), and Gemini CLI integration, all designed to redefine AI-assisted development allow developers to create full-stack applications from a single prompt and integrate powerful AI capabilities directly into their workflow.

    Advancing agentic AI development with Firebase Studio
  • JULY 9, 2025 / Gemma

    T5Gemma: A new collection of encoder-decoder Gemma models

    T5Gemma is a new family of encoder-decoder LLMs developed by converting and adapting pretrained decoder-only models based on the Gemma 2 framework, offering superior performance and efficiency compared to its decoder-only counterparts, particularly for tasks requiring deep input understanding, like summarization and translation.

    T5Gemma: A New Collection of Encoder-Decoder Gemma Models
  • JULY 7, 2025 / Gemini

    Batch Mode in the Gemini API: Process more for less

    The new batch mode in the Gemini API is designed for high-throughput, non-latency-critical AI workloads, simplifying large jobs by handling scheduling and processing, and making tasks like data analysis, bulk content creation, and model evaluation more cost-effective and scalable, so developers can process large volumes of data efficiently.

    Scale your AI workloads with batch mode in the Gemini API
  • JUNE 26, 2025 / Gemma

    Introducing Gemma 3n: The developer guide

    The Gemma 3n model has been fully released, building on the success of previous Gemma models and bringing advanced on-device multimodal capabilities to edge devices with unprecedented performance. Explore Gemma 3n's innovations, including its mobile-first architecture, MatFormer technology, Per-Layer Embeddings, KV Cache Sharing, and new audio and MobileNet-V5 vision encoders, and how developers can start building with it today.

    Introducing Gemma 3n: The Developer Guide