Skip to content
Open
Show file tree
Hide file tree
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion examples/Animated_Story_Video_Generation_gemini.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -278,7 +278,7 @@
" 'api_version': 'v1alpha'\n",
"})\n",
"# Create a client for text generation using Gemini.\n",
"MODEL = \"gemini-3.1-flash-lite-preview\"\n",
"MODEL = \"gemini-3-flash-preview\"\n",
"# Create a client for image generation using Imagen.\n",
"IMAGE_MODEL_ID = \"imagen-3.0-generate-002\"\n"
]
Expand Down
6 changes: 3 additions & 3 deletions examples/Anomaly_detection_with_embeddings.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -71,8 +71,8 @@
"name": "stdout",
"output_type": "stream",
"text": [
"\u001b[?25l \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m0.0/137.7 kB\u001b[0m \u001b[31m?\u001b[0m eta \u001b[36m-:--:--\u001b[0m\r",
"\u001b[2K \u001b[91m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[91m╸\u001b[0m\u001b[90m━\u001b[0m \u001b[32m133.1/137.7 kB\u001b[0m \u001b[31m4.6 MB/s\u001b[0m eta \u001b[36m0:00:01\u001b[0m\r",
"\u001b[?25l \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m0.0/137.7 kB\u001b[0m \u001b[31m?\u001b[0m eta \u001b[36m-:--:--\u001b[0m\r\n",
Copy link
Copy Markdown
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Why are you adding \n in the output?

"\u001b[2K \u001b[91m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m\u001b[91m╸\u001b[0m\u001b[90m━\u001b[0m \u001b[32m133.1/137.7 kB\u001b[0m \u001b[31m4.6 MB/s\u001b[0m eta \u001b[36m0:00:01\u001b[0m\r\n",
"\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m137.7/137.7 kB\u001b[0m \u001b[31m2.9 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n",
"\u001b[?25h"
]
Expand Down Expand Up @@ -1276,7 +1276,7 @@
"\n",
"\n",
"def create_embeddings(df):\n",
" MODEL_ID = \"text-embedding-004\" # @param [\"embedding-001\",\"text-embedding-004\"] {allow-input: true}\n",
" MODEL_ID = \"text-embedding-001\" # @param [\"embedding-001\",\"text-embedding-004\"] {allow-input: true}\n",
" model = f\"models/{MODEL_ID}\"\n",
" embed_fn = make_embed_text_fn(model)\n",
"\n",
Expand Down
2 changes: 1 addition & 1 deletion examples/Book_illustration.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -207,7 +207,7 @@
"outputs": [],
"source": [
"IMAGE_MODEL_ID = \"gemini-2.5-flash-image\" # @param [\"gemini-2.5-flash-image\", \"gemini-3-pro-image-preview\"] {\"allow-input\":true, isTemplate: true}\n",
"GEMINI_MODEL_ID = \"gemini-2.5-flash\" # @param [\"gemini-2.5-flash-lite\", \"gemini-2.5-flash\", \"gemini-2.5-pro\", \"gemini-3.1-flash-lite-preview\", \"gemini-3-flash-preview\", \"gemini-3.1-pro-preview\"] {\"allow-input\":true, isTemplate: true}"
"GEMINI_MODEL_ID = \"gemini-3-flash-preview\" # @param [\"gemini-2.5-flash-lite\", \"gemini-2.5-flash\", \"gemini-2.5-pro\", \"gemini-3.1-flash-lite-preview\", \"gemini-3-flash-preview\", \"gemini-3.1-pro-preview\"] {\"allow-input\":true, isTemplate: true}"
]
},
{
Expand Down
2 changes: 1 addition & 1 deletion examples/Browser_as_a_tool.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -156,7 +156,7 @@
"client = genai.Client(api_key=GOOGLE_API_KEY)\n",
"\n",
"LIVE_MODEL = 'gemini-2.5-flash-native-audio-preview-09-2025' # @param ['gemini-2.0-flash-live-001', 'gemini-live-2.5-flash-preview', 'gemini-2.5-flash-native-audio-preview-09-2025'] {allow-input: true, isTemplate: true}\n",
"MODEL = 'gemini-2.5-flash' # @param ['gemini-2.5-flash'] {allow-input: true, isTemplate: true}"
"MODEL = 'gemini-3-flash-preview' # @param ['gemini-2.5-flash'] {allow-input: true, isTemplate: true}"
]
},
{
Expand Down
2 changes: 1 addition & 1 deletion examples/Datasets.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -248,7 +248,7 @@
}
],
"source": [
"models_to_eval = ['gemini-2.5-flash', 'gemini-2.5-pro']\n",
"models_to_eval = ['gemini-3-flash-preview', 'gemini-3-pro-preview']\n",
"\n",
"batches = []\n",
"for model in models_to_eval:\n",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -91,7 +91,7 @@
"client = genai.Client(api_key=api_key)\n",
"\n",
"# Define the model you are going to use\n",
"model_id = \"gemini-2.5-flash\" # or \"gemini-2.5-flash-lite\", \"gemini-2.5-flash\", \"gemini-2.5-pro\", \"gemini-3.1-flash-lite-preview\", \"gemini-3.1-pro-preview\""
"model_id = \"gemini-3-flash-preview\" # or \"gemini-2.5-flash-lite\", \"gemini-2.5-flash\", \"gemini-2.5-pro\", \"gemini-3.1-flash-lite-preview\", \"gemini-3.1-pro-preview\""
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

The comment listing alternative models is outdated and doesn't include the new default gemini-3-flash-preview. According to the repository style guide, it's recommended to use a Colab parameter selector for model selection to improve maintainability. Please consider converting this to a @param dropdown.

model_id = "gemini-3-flash-preview" # @param ["gemini-3-flash-preview", "gemini-3-pro-preview", "gemini-3.1-flash-lite-preview", "gemini-3.1-pro-preview"] {"allow-input":true, isTemplate: true}
References
  1. When selecting a model, it is recommended to use a Colab parameter selector (@param) to make it easier for users to switch between models and for maintainers to update them. The list of models should be kept consistent and up-to-date. (link)

]
},
{
Expand Down
2 changes: 1 addition & 1 deletion examples/Search_Wikipedia_using_ReAct.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -854,7 +854,7 @@
}
],
"source": [
"gemini_ReAct_chat = ReAct(model='gemini-2.0-flash', ReAct_prompt='model_instructions.txt')\n",
"gemini_ReAct_chat = ReAct(model='gemini-3-flash-preview', ReAct_prompt='model_instructions.txt')\n",
"# Note: try different combinations of generational_config parameters for variational results\n",
"gemini_ReAct_chat(\"What are the total of ages of the main trio from the new Percy Jackson and the Olympians TV series in real life?\", temperature=0.2)"
]
Expand Down
2 changes: 1 addition & 1 deletion examples/Spatial_understanding_3d.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -2227,7 +2227,7 @@
}
],
"source": [
"PRO_MODEL_ID ='gemini-2.5-pro'\n",
"PRO_MODEL_ID ='gemini-3-flash-preview'\n",
"\n",
"# Load and resize the image.\n",
"img_0 = Image.open(\"music_0.jpg\")\n",
Expand Down
2 changes: 1 addition & 1 deletion examples/Tag_and_caption_images.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -230,7 +230,7 @@
"from PIL import Image as PILImage\n",
"import time\n",
"\n",
"MODEL_ID='gemini-2.0-flash' # @param [\"gemini-2.5-flash-lite\", \"gemini-2.5-flash\", \"gemini-2.5-pro\", \"gemini-3.1-flash-lite-preview\", \"gemini-3.1-pro-preview\"] {\"allow-input\":true, isTemplate: true}\n",
"MODEL_ID='gemini-3-flash-preview' # @param [\"gemini-2.5-flash-lite\", \"gemini-2.5-flash\", \"gemini-2.5-pro\", \"gemini-3.1-flash-lite-preview\", \"gemini-3.1-pro-preview\"] {\"allow-input\":true, isTemplate: true}\n",
"\n",
"# a helper function for calling\n",
"\n",
Expand Down
2 changes: 1 addition & 1 deletion examples/Voice_memos.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -325,7 +325,7 @@
"\n",
"prompt = \"Draft my next blog post based on my thoughts in this audio file and these two previous blog posts I wrote.\"\n",
"\n",
"MODEL_ID =\"gemini-2.5-flash\" # @param [\"gemini-2.5-flash\", \"gemini-2.5-pro\"] {\"allow-input\":true, isTemplate: true}\n",
"MODEL_ID =\"gemini-3-flash-preview\" # @param [\"gemini-2.5-flash\", \"gemini-2.5-pro\"] {\"allow-input\":true, isTemplate: true}\n",
"\n",
"response = client.models.generate_content(\n",
" model=MODEL_ID,\n",
Expand Down
2 changes: 1 addition & 1 deletion examples/fastrtc_ui.py
Original file line number Diff line number Diff line change
Expand Up @@ -113,7 +113,7 @@ async def start_up(self):
),
)
async with client.aio.live.connect(
model="gemini-2.5-flash-lite", config=config
model="gemini-3-flash-preview", config=config
) as session:
async for audio in session.start_stream(
stream=self.stream(), mime_type="audio/pcm"
Expand Down
2 changes: 1 addition & 1 deletion examples/gemini_google_adk_model_guardrails.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -330,7 +330,7 @@
"source": [
"# Create an UNSAFE agent (no guardrails)\n",
"unsafe_agent = llm_agent.LlmAgent(\n",
" model=\"gemini-2.5-flash\",\n",
" model=\"gemini-3-flash-preview\",\n",
" name=\"unsafe_calculator_agent\",\n",
" instruction=\"\"\"You are a helpful calculator assistant.\n",
" Help users with mathematical calculations.\"\"\",\n",
Expand Down
51 changes: 47 additions & 4 deletions examples/langchain/Chat_with_SQL_using_langchain.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -210,7 +210,28 @@
"outputs": [
{
"data": {
"text/markdown": "\nCREATE TABLE housing (\n\t\"MedInc\" REAL, \n\t\"HouseAge\" REAL, \n\t\"AveRooms\" REAL, \n\t\"AveBedrms\" REAL, \n\t\"Population\" REAL, \n\t\"AveOccup\" REAL, \n\t\"Latitude\" REAL, \n\t\"Longitude\" REAL, \n\t\"MedHouseVal\" REAL\n)\n\n/*\n3 rows from housing table:\nMedInc\tHouseAge\tAveRooms\tAveBedrms\tPopulation\tAveOccup\tLatitude\tLongitude\tMedHouseVal\n8.3252\t41.0\t6.984126984126984\t1.0238095238095237\t322.0\t2.5555555555555554\t37.88\t-122.23\t4.526\n8.3014\t21.0\t6.238137082601054\t0.9718804920913884\t2401.0\t2.109841827768014\t37.86\t-122.22\t3.585\n7.2574\t52.0\t8.288135593220339\t1.073446327683616\t496.0\t2.8022598870056497\t37.85\t-122.24\t3.521\n*/",
"text/markdown": [
"\n",
"CREATE TABLE housing (\n",
"\t\"MedInc\" REAL, \n",
"\t\"HouseAge\" REAL, \n",
"\t\"AveRooms\" REAL, \n",
"\t\"AveBedrms\" REAL, \n",
"\t\"Population\" REAL, \n",
"\t\"AveOccup\" REAL, \n",
"\t\"Latitude\" REAL, \n",
"\t\"Longitude\" REAL, \n",
"\t\"MedHouseVal\" REAL\n",
")\n",
"\n",
"/*\n",
"3 rows from housing table:\n",
"MedInc\tHouseAge\tAveRooms\tAveBedrms\tPopulation\tAveOccup\tLatitude\tLongitude\tMedHouseVal\n",
"8.3252\t41.0\t6.984126984126984\t1.0238095238095237\t322.0\t2.5555555555555554\t37.88\t-122.23\t4.526\n",
"8.3014\t21.0\t6.238137082601054\t0.9718804920913884\t2401.0\t2.109841827768014\t37.86\t-122.22\t3.585\n",
"7.2574\t52.0\t8.288135593220339\t1.073446327683616\t496.0\t2.8022598870056497\t37.85\t-122.24\t3.521\n",
"*/"
],
"text/plain": [
"<IPython.core.display.Markdown object>"
]
Expand All @@ -234,7 +255,7 @@
"outputs": [],
"source": [
"# Define query chain\n",
"llm = ChatGoogleGenerativeAI(model=\"gemini-2.5-flash\", temperature=0)\n",
"llm = ChatGoogleGenerativeAI(model=\"gemini-3-flash-preview\", temperature=0)\n",
"write_query_chain = create_sql_query_chain(llm, db)"
]
},
Expand All @@ -259,7 +280,25 @@
"outputs": [
{
"data": {
"text/markdown": "You are a SQLite expert. Given an input question, first create a syntactically correct SQLite query to run, then look at the results of the query and return the answer to the input question.\nUnless the user specifies in the question a specific number of examples to obtain, query for at most {top_k} results using the LIMIT clause as per SQLite. You can order the results to return the most informative data in the database.\nNever query for all columns from a table. You must query only the columns that are needed to answer the question. Wrap each column name in double quotes (\") to denote them as delimited identifiers.\nPay attention to use only the column names you can see in the tables below. Be careful to not query for columns that do not exist. Also, pay attention to which column is in which table.\nPay attention to use date('now') function to get the current date, if the question involves \"today\".\n\nUse the following format:\n\nQuestion: Question here\nSQLQuery: SQL Query to run\nSQLResult: Result of the SQLQuery\nAnswer: Final answer here\n\nOnly use the following tables:\n{table_info}\n\nQuestion: {input}",
"text/markdown": [
"You are a SQLite expert. Given an input question, first create a syntactically correct SQLite query to run, then look at the results of the query and return the answer to the input question.\n",
"Unless the user specifies in the question a specific number of examples to obtain, query for at most {top_k} results using the LIMIT clause as per SQLite. You can order the results to return the most informative data in the database.\n",
"Never query for all columns from a table. You must query only the columns that are needed to answer the question. Wrap each column name in double quotes (\") to denote them as delimited identifiers.\n",
"Pay attention to use only the column names you can see in the tables below. Be careful to not query for columns that do not exist. Also, pay attention to which column is in which table.\n",
"Pay attention to use date('now') function to get the current date, if the question involves \"today\".\n",
"\n",
"Use the following format:\n",
"\n",
"Question: Question here\n",
"SQLQuery: SQL Query to run\n",
"SQLResult: Result of the SQLQuery\n",
"Answer: Final answer here\n",
"\n",
"Only use the following tables:\n",
"{table_info}\n",
"\n",
"Question: {input}"
],
"text/plain": [
"<IPython.core.display.Markdown object>"
]
Expand All @@ -282,7 +321,11 @@
"outputs": [
{
"data": {
"text/markdown": "```sqlite\nSELECT sum(\"Population\") FROM housing\n```",
"text/markdown": [
"```sqlite\n",
"SELECT sum(\"Population\") FROM housing\n",
"```"
],
"text/plain": [
"<IPython.core.display.Markdown object>"
]
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -416,7 +416,7 @@
"# If you only want to set a custom temperature for the model use the\n",
"# \"temperature\" parameter directly.\n",
"\n",
"llm = ChatGoogleGenerativeAI(model=\"gemini-2.5-flash\")"
"llm = ChatGoogleGenerativeAI(model=\"gemini-3-flash-preview\")"
]
},
{
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -491,7 +491,7 @@
"# If you only want to set a custom temperature for the model use the\n",
"# \"temperature\" parameter directly.\n",
"\n",
"llm = ChatGoogleGenerativeAI(model=\"gemini-2.5-flash\")"
"llm = ChatGoogleGenerativeAI(model=\"gemini-3-flash-preview\")"
]
},
{
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -243,7 +243,7 @@
"# If you only want to set a custom temperature for the model use the\n",
"# \"temperature\" parameter directly.\n",
"\n",
"llm = ChatGoogleGenerativeAI(model=\"gemini-2.5-flash\")"
"llm = ChatGoogleGenerativeAI(model=\"gemini-3-flash-preview\")"
]
},
{
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -329,7 +329,7 @@
"# If you only want to set a custom temperature for the model use the\n",
"# \"temperature\" parameter directly.\n",
"\n",
"llm = GoogleGenAI(model_name=\"models/gemini-2.5-flash\")"
"llm = GoogleGenAI(model_name=\"models/gemini-3-flash-preview\")"
]
},
{
Expand Down
10 changes: 7 additions & 3 deletions examples/prompting/Basic_Classification.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -174,7 +174,9 @@
"outputs": [
{
"data": {
"text/markdown": "Spam\n",
"text/markdown": [
"Spam\n"
],
"text/plain": [
"<IPython.core.display.Markdown object>"
]
Expand All @@ -198,7 +200,7 @@
")\n",
"\n",
"response = client.models.generate_content(\n",
" model='gemini-2.5-flash',\n",
" model='gemini-3-flash-preview',\n",
" contents=spam_prompt,\n",
" config=generation_config\n",
")\n",
Expand All @@ -214,7 +216,9 @@
"outputs": [
{
"data": {
"text/markdown": "Neutral\n",
"text/markdown": [
"Neutral\n"
],
"text/plain": [
"<IPython.core.display.Markdown object>"
]
Expand Down
2 changes: 1 addition & 1 deletion examples/prompting/Basic_Evaluation.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -160,7 +160,7 @@
" Ensure your essay includes a clear thesis statement. You should write only an essay, so do not include any notes.\n",
"\"\"\"\n",
"\n",
"MODEL_ID=\"gemini-2.5-flash\" # @param [\"gemini-2.5-flash-lite\",\"gemini-2.5-flash\",\"gemini-2.5-pro\", \"gemini-3.1-flash-lite-preview\", \"gemini-3.1-pro-preview\"] {\"allow-input\":true, isTemplate: true}\n",
"MODEL_ID=\"gemini-3-flash-preview\" # @param [\"gemini-2.5-flash-lite\",\"gemini-2.5-flash\",\"gemini-2.5-pro\", \"gemini-3.1-flash-lite-preview\", \"gemini-3.1-pro-preview\"] {\"allow-input\":true, isTemplate: true}\n",
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

The default MODEL_ID has been updated to gemini-3-flash-preview, but this model is not in the @param list. The list also contains several deprecated models. To ensure the Colab UI works as expected and to align with the goal of this PR, please update the list of models.

MODEL_ID="gemini-3-flash-preview" # @param ["gemini-3-flash-preview", "gemini-3-pro-preview", "gemini-3.1-flash-lite-preview", "gemini-3.1-pro-preview"] {"allow-input":true, isTemplate: true}

References
  1. When selecting a model, it is recommended to use a Colab parameter selector (@param) to make it easier for users to switch between models and for maintainers to update them. The list of models should be kept consistent and up-to-date. (link)

"\n",
"student_response = client.models.generate_content(\n",
" model=MODEL_ID,\n",
Expand Down
2 changes: 1 addition & 1 deletion examples/prompting/Few_shot_prompting.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -125,7 +125,7 @@
},
"outputs": [],
"source": [
"MODEL_ID=\"gemini-2.5-flash\" # @param [\"gemini-2.5-flash-lite\",\"gemini-2.5-flash\",\"gemini-2.5-pro\",\"gemini-3.1-flash-lite-preview\",\"gemini-3.1-pro-preview\"] {\"allow-input\":true, isTemplate: true}"
"MODEL_ID=\"gemini-3-flash-preview\" # @param [\"gemini-2.5-flash-lite\",\"gemini-2.5-flash\",\"gemini-2.5-pro\",\"gemini-3.1-flash-lite-preview\",\"gemini-3.1-pro-preview\"] {\"allow-input\":true, isTemplate: true}"
]
},
{
Expand Down
16 changes: 11 additions & 5 deletions examples/prompting/Providing_base_cases.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -213,7 +213,7 @@
},
"outputs": [],
"source": [
"model = genai.GenerativeModel(model_name='gemini-2.0-flash', system_instruction=instructions)"
"model = genai.GenerativeModel(model_name='gemini-3-flash-preview', system_instruction=instructions)"
]
},
{
Expand All @@ -227,7 +227,8 @@
"name": "stdout",
"output_type": "stream",
"text": [
"## Specified genre:\n\n",
"## Specified genre:\n",
"\n",
"Of course! I'd be happy to recommend some books with hard magic systems. \n",
"\n",
"Here are three recommendations:\n",
Expand All @@ -240,7 +241,8 @@
"\n",
"Enjoy your reading! \n",
"\n",
"## Not specified genre:\n\n",
"## Not specified genre:\n",
"\n",
"Sure! Since you didn't specify a genre, I'll recommend two spine-chilling horror novels:\n",
"\n",
"1. **\"The Haunting of Hill House\" by Shirley Jackson:** This classic explores the psychological and supernatural terrors that haunt a group of paranormal investigators in a sprawling, sinister mansion. \n",
Expand All @@ -253,8 +255,12 @@
}
],
"source": [
"print(\"## Specified genre:\n\n\", model.generate_content(\"Could you recommend me 3 books with hard magic system?\").text, sep=\"\\n\")\n",
"print(\"## Not specified genre:\n\n\", model.generate_content(\"Could you recommend me 2 books?\").text, sep=\"\\n\")"
"print(\"## Specified genre:\n",
"\n",
"\", model.generate_content(\"Could you recommend me 3 books with hard magic system?\").text, sep=\"\\n\")\n",
"print(\"## Not specified genre:\n",
"\n",
"\", model.generate_content(\"Could you recommend me 2 books?\").text, sep=\"\\n\")"
]
},
{
Expand Down
2 changes: 1 addition & 1 deletion examples/prompting/Role_prompting.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -143,7 +143,7 @@
"source": [
"prompt = 'Write a 2 paragraph long review of Requiem.'\n",
"\n",
"MODEL_ID=\"gemini-2.5-flash\" # @param [\"gemini-2.5-flash-lite\",\"gemini-2.5-flash\",\"gemini-2.5-pro\",\"gemini-3.1-flash-lite-preview\",\"gemini-3.1-pro-preview\"] {\"allow-input\":true, isTemplate: true}\n",
"MODEL_ID=\"gemini-3-flash-preview\" # @param [\"gemini-2.5-flash-lite\",\"gemini-2.5-flash\",\"gemini-2.5-pro\",\"gemini-3.1-flash-lite-preview\",\"gemini-3.1-pro-preview\"] {\"allow-input\":true, isTemplate: true}\n",
"\n",
"response = client.models.generate_content(\n",
" model=MODEL_ID,\n",
Expand Down
2 changes: 1 addition & 1 deletion examples/prompting/Zero_shot_prompting.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -123,7 +123,7 @@
},
"outputs": [],
"source": [
"MODEL_ID=\"gemini-2.5-flash\" # @param [\"gemini-2.5-flash-lite\",\"gemini-2.5-flash\",\"gemini-2.5-pro\",\"gemini-3.1-flash-lite-preview\",\"gemini-3.1-pro-preview\"] {\"allow-input\":true, isTemplate: true}"
"MODEL_ID=\"gemini-3-flash-preview\" # @param [\"gemini-2.5-flash-lite\",\"gemini-2.5-flash\",\"gemini-2.5-pro\",\"gemini-3.1-flash-lite-preview\",\"gemini-3.1-pro-preview\"] {\"allow-input\":true, isTemplate: true}"
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

The default MODEL_ID has been updated to gemini-3-flash-preview, but this model is not in the @param list. The list also contains several deprecated models. To ensure the Colab UI works as expected and to align with the goal of this PR, please update the list of models.

MODEL_ID="gemini-3-flash-preview" # @param ["gemini-3-flash-preview", "gemini-3-pro-preview", "gemini-3.1-flash-lite-preview", "gemini-3.1-pro-preview"] {"allow-input":true, isTemplate: true}
References
  1. When selecting a model, it is recommended to use a Colab parameter selector (@param) to make it easier for users to switch between models and for maintainers to update them. The list of models should be kept consistent and up-to-date. (link)

]
},
{
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -359,7 +359,7 @@
"source": [
"PROJECT_ID = \"\" # leave this empty\n",
"API_ENDPOINT = \"generativelanguage.googleapis.com\"\n",
"embedding_model = \"text-embedding-004\" # embedding model\n",
"embedding_model = \"text-embedding-001\" # embedding model\n",
"generative_model = \"gemini-3-flash-preview\" # language model\n",
"\n",
"# Products Collection\n",
Expand Down
Loading
Loading