From 2833e9762554ffd2a46617073cbca78356398c56 Mon Sep 17 00:00:00 2001 From: tronikos Date: Mon, 23 Jun 2025 11:11:16 -0700 Subject: [PATCH] Default to gemini-2.5-flash (#147334) --- .../google_generative_ai_conversation/const.py | 2 +- .../snapshots/test_diagnostics.ambr | 2 +- .../snapshots/test_init.ambr | 8 ++++---- .../test_config_flow.py | 14 +++++++------- 4 files changed, 13 insertions(+), 13 deletions(-) diff --git a/homeassistant/components/google_generative_ai_conversation/const.py b/homeassistant/components/google_generative_ai_conversation/const.py index 831e7d8f508..7e699d7c8c0 100644 --- a/homeassistant/components/google_generative_ai_conversation/const.py +++ b/homeassistant/components/google_generative_ai_conversation/const.py @@ -9,7 +9,7 @@ CONF_PROMPT = "prompt" ATTR_MODEL = "model" CONF_RECOMMENDED = "recommended" CONF_CHAT_MODEL = "chat_model" -RECOMMENDED_CHAT_MODEL = "models/gemini-2.0-flash" +RECOMMENDED_CHAT_MODEL = "models/gemini-2.5-flash" RECOMMENDED_TTS_MODEL = "gemini-2.5-flash-preview-tts" CONF_TEMPERATURE = "temperature" RECOMMENDED_TEMPERATURE = 1.0 diff --git a/tests/components/google_generative_ai_conversation/snapshots/test_diagnostics.ambr b/tests/components/google_generative_ai_conversation/snapshots/test_diagnostics.ambr index 60d388d0502..a31827c7acc 100644 --- a/tests/components/google_generative_ai_conversation/snapshots/test_diagnostics.ambr +++ b/tests/components/google_generative_ai_conversation/snapshots/test_diagnostics.ambr @@ -5,7 +5,7 @@ 'api_key': '**REDACTED**', }), 'options': dict({ - 'chat_model': 'models/gemini-2.0-flash', + 'chat_model': 'models/gemini-2.5-flash', 'dangerous_block_threshold': 'BLOCK_MEDIUM_AND_ABOVE', 'harassment_block_threshold': 'BLOCK_MEDIUM_AND_ABOVE', 'hate_block_threshold': 'BLOCK_MEDIUM_AND_ABOVE', diff --git a/tests/components/google_generative_ai_conversation/snapshots/test_init.ambr b/tests/components/google_generative_ai_conversation/snapshots/test_init.ambr index d8e54b15f61..f89871ff131 100644 --- a/tests/components/google_generative_ai_conversation/snapshots/test_init.ambr +++ b/tests/components/google_generative_ai_conversation/snapshots/test_init.ambr @@ -11,7 +11,7 @@ File(name='doorbell_snapshot.jpg', display_name=None, mime_type=None, size_bytes=None, create_time=None, expiration_time=None, update_time=None, sha256_hash=None, uri=None, download_uri=None, state=, source=None, video_metadata=None, error=None), File(name='context.txt', display_name=None, mime_type=None, size_bytes=None, create_time=None, expiration_time=None, update_time=None, sha256_hash=None, uri=None, download_uri=None, state=, source=None, video_metadata=None, error=None), ]), - 'model': 'models/gemini-2.0-flash', + 'model': 'models/gemini-2.5-flash', }), ), ]) @@ -28,7 +28,7 @@ b'some file', b'some file', ]), - 'model': 'models/gemini-2.0-flash', + 'model': 'models/gemini-2.5-flash', }), ), ]) @@ -43,7 +43,7 @@ 'contents': list([ 'Write an opening speech for a Home Assistant release party', ]), - 'model': 'models/gemini-2.0-flash', + 'model': 'models/gemini-2.5-flash', }), ), ]) @@ -58,7 +58,7 @@ 'contents': list([ 'Write an opening speech for a Home Assistant release party', ]), - 'model': 'models/gemini-2.0-flash', + 'model': 'models/gemini-2.5-flash', }), ), ]) diff --git a/tests/components/google_generative_ai_conversation/test_config_flow.py b/tests/components/google_generative_ai_conversation/test_config_flow.py index 4234355cb5b..0dc0996ad30 100644 --- a/tests/components/google_generative_ai_conversation/test_config_flow.py +++ b/tests/components/google_generative_ai_conversation/test_config_flow.py @@ -41,6 +41,12 @@ from tests.common import MockConfigEntry def get_models_pager(): """Return a generator that yields the models.""" + model_25_flash = Mock( + display_name="Gemini 2.5 Flash", + supported_actions=["generateContent"], + ) + model_25_flash.name = "models/gemini-2.5-flash" + model_20_flash = Mock( display_name="Gemini 2.0 Flash", supported_actions=["generateContent"], @@ -59,17 +65,11 @@ def get_models_pager(): ) model_15_pro.name = "models/gemini-1.5-pro-latest" - model_10_pro = Mock( - display_name="Gemini 1.0 Pro", - supported_actions=["generateContent"], - ) - model_10_pro.name = "models/gemini-pro" - async def models_pager(): + yield model_25_flash yield model_20_flash yield model_15_flash yield model_15_pro - yield model_10_pro return models_pager()