Skip to content

Conversation

@ilopezluna
Copy link
Contributor

No description provided.

@ilopezluna ilopezluna requested a review from a team January 2, 2026 10:39
@gemini-code-assist
Copy link
Contributor

Summary of Changes

Hello @ilopezluna, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed!

This pull request expands the model card collection by introducing documentation for the FunctionGemma model in two variants: one for vLLM and another for the GGUF version by Unsloth. These additions provide detailed information on the model's architecture, intended applications, and integration with Docker Model Runner, enhancing the discoverability and usability of this specialized function-calling AI model.

Highlights

  • New Model Card: FunctionGemma (vLLM): A new model card has been added for FunctionGemma, specifically tailored for vLLM. This card provides a comprehensive overview of the model, including its intended uses, input/output specifications, and instructions for running it with Docker Model Runner.
  • New Model Card: FunctionGemma (GGUF by Unsloth): Another new model card has been introduced for the GGUF version of FunctionGemma, optimized by Unsloth. Similar to the vLLM version, this card details the model's capabilities, use cases, and Docker integration, with an additional link to Unsloth's GGUF documentation.
  • FunctionGemma Model Details: Both model cards describe FunctionGemma as a lightweight 270M-parameter open model from Google, built on Gemma 3 and trained for text-only function calling. It's designed for fine-tuning into efficient, offline-capable specialized agents for resource-constrained devices.

🧠 New Feature in Public Preview: You can now enable Memory to help Gemini Code Assist learn from your team's feedback. This makes future code reviews more consistent and personalized to your project's style. Click here to enable Memory in your admin console.

Using Gemini Code Assist

The full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips.

Invoking Gemini

You can request assistance from Gemini at any point by creating a comment using either /gemini <command> or @gemini-code-assist <command>. Below is a summary of the supported commands on the current page.

Feature Command Description
Code Review /gemini review Performs a code review for the current pull request in its current state.
Pull Request Summary /gemini summary Provides a summary of the current pull request in its current state.
Comment @gemini-code-assist Responds in comments when explicitly tagged, both in pull request comments and review comments.
Help /gemini help Displays a list of available commands.

Customization

To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a .gemini/ folder in the base of the repository. Detailed instructions can be found here.

Limitations & Feedback

Gemini Code Assist may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments. If you're interested in giving your feedback about your experience with Gemini Code Assist for Github and other Google products, sign up here.

You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension.

Footnotes

  1. Review the Privacy Notices, Generative AI Prohibited Use Policy, Terms of Service, and learn how to configure Gemini Code Assist in GitHub here. Gemini can make mistakes, so double check it and use code with caution.

Copy link
Contributor

@sourcery-ai sourcery-ai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Hey - I've left some high level feedback:

  • The two model card files duplicate the same long-form description; consider extracting shared content or clearly highlighting only the differences (e.g., GGUF vs vLLM) to make maintenance easier and reduce the risk of divergence.
  • Only the GGUF variant mentions Unsloth explicitly and links to the Unsloth GGUF docs, while the vLLM card omits any vLLM-specific link or explanation; aligning both cards so each clearly calls out its serving/runtime specifics would make the distinction between them clearer.
Prompt for AI Agents
Please address the comments from this code review:

## Overall Comments
- The two model card files duplicate the same long-form description; consider extracting shared content or clearly highlighting only the differences (e.g., GGUF vs vLLM) to make maintenance easier and reduce the risk of divergence.
- Only the GGUF variant mentions Unsloth explicitly and links to the Unsloth GGUF docs, while the vLLM card omits any vLLM-specific link or explanation; aligning both cards so each clearly calls out its serving/runtime specifics would make the distinction between them clearer.

Sourcery is free for open source - if you like our reviews please consider sharing them ✨
Help me be more useful! Please click 👍 or 👎 on each comment and I'll use the feedback to improve your reviews.

Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request adds model cards for two variants of FunctionGemma: a vLLM version and a GGUF version. The cards provide a good overview of the model. However, I've found a few critical and high-severity issues that need to be addressed.

  • There's a factual error regarding the model's context window, which is 8K tokens, not 32K as stated.
  • The base model is incorrectly cited as 'Gemma 3' instead of 'Gemma 2'.
  • There are also some minor typos and opportunities to improve readability by hyphenating compound words and breaking up a very long paragraph.
    I've provided specific suggestions to correct these issues in both files. Addressing these will ensure the model cards are accurate and easy to read.

## Inputs and outputs
Input:
- Text string, such as a question, a prompt, or a document to be summarized
- Total input context of 32K tokens
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

critical

The context window for FunctionGemma is 8K tokens, not 32K. Please correct this value. The official documentation states an 8,192 token context window.

Suggested change
- Total input context of 32K tokens
- Total input context of 8K tokens


Output:
- Generated text in response to the input, such as an answer to a question, or a summary of a document
- Total output context up to 32K tokens per request, subtracting the request input tokens
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

critical

The total output context should also reflect the 8K token limit. The maximum output is 8K minus the input tokens.

Suggested change
- Total output context up to 32K tokens per request, subtracting the request input tokens
- Total output context up to 8K tokens per request, subtracting the request input tokens

## Inputs and outputs
Input:
- Text string, such as a question, a prompt, or a document to be summarized
- Total input context of 32K tokens
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

critical

The context window for FunctionGemma is 8K tokens, not 32K. Please correct this value to align with the official model specifications, which state an 8,192 token context window.

Suggested change
- Total input context of 32K tokens
- Total input context of 8K tokens


Output:
- Generated text in response to the input, such as an answer to a question, or a summary of a document
- Total output context up to 32K tokens per request, subtracting the request input tokens
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

critical

The total output context should also reflect the 8K token limit. The maximum output is 8K minus the number of input tokens.

Suggested change
- Total output context up to 32K tokens per request, subtracting the request input tokens
- Total output context up to 8K tokens per request, subtracting the request input tokens


## Intended uses

FunctionGemma is a lightweight, open model from Google, built as a foundation for creating your own specialized function calling models. FunctionGemma is not intended for use as a direct dialogue model, and is designed to be highly performant after further fine-tuning, as is typical of models this size. Built on the Gemma 3 270M model and with the same research and technology used to create the Gemini models, FunctionGemma has been trained specifically for function calling. The model has the same architecture as Gemma 3, but uses a different chat format. The model is well suited for text-only function calling. The uniquely small size makes it possible to deploy in environments with limited resources such as laptops, desktops or your own cloud infrastructure, democratizing access to state of the art AI models and helping foster innovation for everyone. Furthermore, akin to the base Gemma 270M, the model has been optimized to be extremely versatile, performant on a variety of hardware in single turn scenarios, but should be finetuned on single turn or multiturn task specific data to achieve best accuracy in specific domains. To demonstrate how specializing the 270M parameter model can achieve high performance on specific agentic workflows, we have highlighted two use cases in the Google AI Edge Gallery app.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

high

This paragraph has some opportunities for improvement in accuracy, consistency, and readability.

  • Accuracy: FunctionGemma is based on Gemma 2, not Gemma 3. This is mentioned twice and should be corrected.
  • Typos: 'state of a art' should be 'state of the art'.
  • Consistency: For better readability, terms like 'finetuned', 'single turn', and 'multiturn' should be hyphenated ('fine-tuned', 'single-turn', 'multi-turn').
  • Readability: The paragraph is very long. Consider splitting it into smaller paragraphs to make it easier to digest.

I've provided a code suggestion that fixes the accuracy and typo issues. You may want to further break up the paragraph for readability.

Suggested change
FunctionGemma is a lightweight, open model from Google, built as a foundation for creating your own specialized function calling models. FunctionGemma is not intended for use as a direct dialogue model, and is designed to be highly performant after further fine-tuning, as is typical of models this size. Built on the Gemma 3 270M model and with the same research and technology used to create the Gemini models, FunctionGemma has been trained specifically for function calling. The model has the same architecture as Gemma 3, but uses a different chat format. The model is well suited for text-only function calling. The uniquely small size makes it possible to deploy in environments with limited resources such as laptops, desktops or your own cloud infrastructure, democratizing access to state of the art AI models and helping foster innovation for everyone. Furthermore, akin to the base Gemma 270M, the model has been optimized to be extremely versatile, performant on a variety of hardware in single turn scenarios, but should be finetuned on single turn or multiturn task specific data to achieve best accuracy in specific domains. To demonstrate how specializing the 270M parameter model can achieve high performance on specific agentic workflows, we have highlighted two use cases in the Google AI Edge Gallery app.
FunctionGemma is a lightweight, open model from Google, built as a foundation for creating your own specialized function calling models. FunctionGemma is not intended for use as a direct dialogue model, and is designed to be highly performant after further fine-tuning, as is typical of models this size. Built on the Gemma 2 270M model and with the same research and technology used to create the Gemini models, FunctionGemma has been trained specifically for function calling. The model has the same architecture as Gemma 2, but uses a different chat format. The model is well suited for text-only function calling. The uniquely small size makes it possible to deploy in environments with limited resources such as laptops, desktops or your own cloud infrastructure, democratizing access to state of the art AI models and helping foster innovation for everyone. Furthermore, akin to the base Gemma 270M, the model has been optimized to be extremely versatile, performant on a variety of hardware in single turn scenarios, but should be fine-tuned on single-turn or multi-turn task specific data to achieve best accuracy in specific domains. To demonstrate how specializing the 270M parameter model can achieve high performance on specific agentic workflows, we have highlighted two use cases in the Google AI Edge Gallery app.


## Intended uses

FunctionGemma is a lightweight, open model from Google, built as a foundation for creating your own specialized function calling models. FunctionGemma is not intended for use as a direct dialogue model, and is designed to be highly performant after further fine-tuning, as is typical of models this size. Built on the Gemma 3 270M model and with the same research and technology used to create the Gemini models, FunctionGemma has been trained specifically for function calling. The model has the same architecture as Gemma 3, but uses a different chat format. The model is well suited for text-only function calling. The uniquely small size makes it possible to deploy in environments with limited resources such as laptops, desktops or your own cloud infrastructure, democratizing access to state of the art AI models and helping foster innovation for everyone. Furthermore, akin to the base Gemma 270M, the model has been optimized to be extremely versatile, performant on a variety of hardware in single turn scenarios, but should be finetuned on single turn or multiturn task specific data to achieve best accuracy in specific domains. To demonstrate how specializing the 270M parameter model can achieve high performance on specific agentic workflows, we have highlighted two use cases in the Google AI Edge Gallery app.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

high

This paragraph has some opportunities for improvement in accuracy, consistency, and readability.

  • Accuracy: FunctionGemma is based on Gemma 2, not Gemma 3. This is mentioned twice and should be corrected.
  • Typos: 'state of a art' should be 'state of the art'.
  • Consistency: For better readability, terms like 'finetuned', 'single turn', and 'multurn' should be hyphenated ('fine-tuned', 'single-turn', 'multi-turn').
  • Readability: The paragraph is very long. Consider splitting it into smaller paragraphs to make it easier to digest.

I've provided a code suggestion that fixes the accuracy and typo issues. You may want to further break up the paragraph for readability.

Suggested change
FunctionGemma is a lightweight, open model from Google, built as a foundation for creating your own specialized function calling models. FunctionGemma is not intended for use as a direct dialogue model, and is designed to be highly performant after further fine-tuning, as is typical of models this size. Built on the Gemma 3 270M model and with the same research and technology used to create the Gemini models, FunctionGemma has been trained specifically for function calling. The model has the same architecture as Gemma 3, but uses a different chat format. The model is well suited for text-only function calling. The uniquely small size makes it possible to deploy in environments with limited resources such as laptops, desktops or your own cloud infrastructure, democratizing access to state of the art AI models and helping foster innovation for everyone. Furthermore, akin to the base Gemma 270M, the model has been optimized to be extremely versatile, performant on a variety of hardware in single turn scenarios, but should be finetuned on single turn or multiturn task specific data to achieve best accuracy in specific domains. To demonstrate how specializing the 270M parameter model can achieve high performance on specific agentic workflows, we have highlighted two use cases in the Google AI Edge Gallery app.
FunctionGemma is a lightweight, open model from Google, built as a foundation for creating your own specialized function calling models. FunctionGemma is not intended for use as a direct dialogue model, and is designed to be highly performant after further fine-tuning, as is typical of models this size. Built on the Gemma 2 270M model and with the same research and technology used to create the Gemini models, FunctionGemma has been trained specifically for function calling. The model has the same architecture as Gemma 2, but uses a different chat format. The model is well suited for text-only function calling. The uniquely small size makes it possible to deploy in environments with limited resources such as laptops, desktops or your own cloud infrastructure, democratizing access to state of the art AI models and helping foster innovation for everyone. Furthermore, akin to the base Gemma 270M, the model has been optimized to be extremely versatile, performant on a variety of hardware in single turn scenarios, but should be fine-tuned on single-turn or multi-turn task specific data to achieve best accuracy in specific domains. To demonstrate how specializing the 270M parameter model can achieve high performance on specific agentic workflows, we have highlighted two use cases in the Google AI Edge Gallery app.

@ilopezluna ilopezluna merged commit 4531817 into main Jan 3, 2026
4 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants