On-Device Function Calling in Google AI Edge Gallery

FEB. 26, 2026
Alice Zheng Product Manager
Milen Ferev Software Engineer
Wai Hon Law Software Engineer

The real magic of AI happens when a model stops merely describing the world and starts interacting with it. One such interaction mechanism is tool-use: the ability to predict and invoke function calls, for example to open apps or adjust system settings.

Shifting tool-use on-device allows developers to build interactions that respond instantly while remaining fully functional regardless of connectivity. This enables, for instance, a natural language voice assistant to instantly create a calendar entry or navigate to a destination while you’re driving.

However, bringing this level of capability to mobile remains a formidable task. Traditional function-calling has historically required large models with memory footprints far exceeding mobile hardware constraints. The real engineering challenge is to compress these models into a mobile footprint, while maintaining accuracy and without draining the battery.

Today, we are excited to announce several major updates to Google’s on-device AI showcase app Google AI Edge Gallery:

  • Building on the cross-platform capabilities of Google AI Edge, we are delighted to bring AI Edge Gallery to iOS in addition to Android, allowing developers to explore the same high-performance, on-device AI use cases powered by Gemma and other open-weight models directly within the iOS ecosystem.
  • We’ve brought the out-of-the-box agentic experiences, Mobile Actions and Tiny Garden, directly to Google AI Edge Gallery, showcasing how Google’s efficient FunctionGemma model translates natural language directly into function calls on device within merely 270M parameters.
  • Building on the state-of-the-art performance benchmarks recently shared in our latest LiteRT announcement, we have now integrated benchmarking as a feature directly into the Google AI Edge Gallery app so you can measure and experience LiteRT’s leading CPU and GPU performance across your own devices.

Experience Mobile Actions and Tiny Garden

The Mobile Actions demo leverages FunctionGemma to reimagine assistant interaction as a fully offline capability. It allows the model to parse natural language commands—such as "Show me the San Francisco airport on map," "Create a calendar event for 2:30 PM tomorrow for cooking class," or "Turn on the flashlight"—and identify the correct OS tool or app intent to execute the command.

The Tiny Garden demo is an interactive mini-game that lets players manage a virtual plot of land using voice commands. For example, a command like "Plant sunflowers in the top row and water them" is decomposed by the model into specific app functions (like plantCrop or waterCrop) targeting grid coordinates. This demonstrates how Google’s compact 270M FunctionGemma model can adapt to highly specific, custom game or app logic directly on a mobile phone, without requiring any server pings.

See Mobile Actions in Action on Android!
You can try this on Android and starting today also on iOS!

Now you've seen these demo experiences in action, you may want to adapt this approach for your own custom use cases. For that, you can fine-tune your own version and implement function calling in the app.

Building on the cross-platform capabilities of Google AI Edge, we are thrilled to bring the full experience of our Android app to the iOS ecosystem with the launch of the Google AI Edge Gallery available in the App Store. Now, iOS developers and enthusiasts can explore the same rich, on-device features, including multi-turn AI Chat, Ask Image queries, and Audio Scribe for local transcription. Most importantly, the iOS app includes our agentic demonstrations, Mobile Actions and Tiny Garden, showcasing how sophisticated tool-calling and function-calling can perform seamlessly on Apple hardware. By leveraging the unified power of the Google AI Edge stack, we’re ensuring that the best of on-device performance, privacy, and offline reliability is accessible to everyone, regardless of their mobile platform.

See Mobile Actions in Action on iOS!

Test model performance in app

Want to see this speed in action? You can now benchmark these models directly within the Gallery app across your own devices (available on Android; coming to iOS soon).

Using Mobile Actions as an example, the performance is blazingly fast on CPU—clocking in at 1916 tokens/sec (prefill) and 142 tokens/sec (decode) on a Pixel 7 Pro.

Here is how to run your own benchmarking tests:

  1. Open the Menu: Tap the hamburger icon in the top-left corner of the Gallery app.
  2. Select Models: Tap the Models tile to see the full list of downloadable models.
  3. Benchmark: Hit the benchmark button and experiment with configurations—adjust prefill/decode tokens or the number of runs—to see exactly how FunctionGemma performs on your specific hardware.
benchmark_wide2
Try it now on Android and see how the model performs on your phone!

Get started today

Ready to build your first local agent? Here is how you can dive in:

  • Explore the demos: Download the Google AI Edge Gallery app (Android and iOS) to see Mobile Actions and Tiny Garden in action.
  • Build your own: Use the fine-tuning recipes to adapt FunctionGemma to your specific app logic and customize AI Edge Gallery with your own functions.
  • Join the conversation: We are constantly iterating. Check out the AI Edge Gallery on GitHub to follow our progress, report issues, or contribute to the future of on-device AI use cases!

We can’t wait to see the agentic features you’ll bring to life. Happy coding!

Acknowledgements

We'd like to extend a special thanks to our key contributors for their foundational work on this project: Francesco Visin, Hriday Chhabria, Jiageng Zhang, Jing Jin, Kat Black, Marissa Ikonomidis, Matthew Chan, Ravin Kumar, Rishika Sinha, Sahil Dua, Xu Chen, Na Li, Yinghao Sun, Yishuang Pang

We also gratefully acknowledge the significant contributions from the following team members: Byungchul Kim, Deepak Nagaraj Halliyavar, Fengwu Yao, Jae Yoo, Jenn Lee, Sahil Dua, Weiyi Wang, Xiaoming Hu, Yasir Modak, Yi-Chun Kuo, Yu-hui Chen, Zhe Chen

This effort was made possible by the guidance and support from our leadership: Cormac Brick, Kathleen Kenealy, Matthias Grundmann, Ram Iyengar , Sachin Kotwani