From c4543408ab67112a2ccad2081db05da71911219e Mon Sep 17 00:00:00 2001 From: kqlio67 <> Date: Thu, 30 Jan 2025 21:11:09 +0200 Subject: [PATCH] Removed model gemma-2-27b because no provider supports this model anymore --- g4f/Provider/Glider.py | 6 +++--- g4f/Provider/needs_auth/HuggingChat.py | 1 - g4f/models.py | 11 +---------- 3 files changed, 4 insertions(+), 14 deletions(-) diff --git a/g4f/Provider/Glider.py b/g4f/Provider/Glider.py index 36472f55a35..8d0e21a71c1 100644 --- a/g4f/Provider/Glider.py +++ b/g4f/Provider/Glider.py @@ -29,9 +29,9 @@ class Glider(AsyncGeneratorProvider, ProviderModelMixin): ] + reasoning_models model_aliases = { - "llama-3-1-70b": "chat-llama-3-1-70b", - "llama-3-1-8b": "chat-llama-3-1-8b", - "llama-3-2-3b": "chat-llama-3-2-3b", + "llama-3.1-70b": "chat-llama-3-1-70b", + "llama-3.1-8b": "chat-llama-3-1-8b", + "llama-3.2-3b": "chat-llama-3-2-3b", "deepseek-r1": "deepseek-ai/DeepSeek-R1", } diff --git a/g4f/Provider/needs_auth/HuggingChat.py b/g4f/Provider/needs_auth/HuggingChat.py index 316a8317b86..bd1c1443412 100644 --- a/g4f/Provider/needs_auth/HuggingChat.py +++ b/g4f/Provider/needs_auth/HuggingChat.py @@ -71,7 +71,6 @@ class HuggingChat(AsyncAuthedProvider, ProviderModelMixin): "flux-schnell": "black-forest-labs/FLUX.1-schnell", ### Used in other providers ### "qwen-2-vl-7b": "Qwen/Qwen2-VL-7B-Instruct", - "gemma-2-27b": "google/gemma-2-27b-it", "qwen-2-72b": "Qwen/Qwen2-72B-Instruct", "qvq-72b": "Qwen/QVQ-72B-Preview", "sd-3.5": "stabilityai/stable-diffusion-3.5-large", diff --git a/g4f/models.py b/g4f/models.py index c20fef8e55c..b27524c6f57 100644 --- a/g4f/models.py +++ b/g4f/models.py @@ -222,7 +222,7 @@ class VisionModel(Model): ) llama_3_2_3b = Model( - name = "llama-3-2-3b", + name = "llama-3.2-3b", base_provider = "Meta Llama", best_provider = Glider ) @@ -347,12 +347,6 @@ class VisionModel(Model): best_provider = Liaobots ) -gemma_2_27b = Model( - name = 'gemma-2-27b', - base_provider = 'Google DeepMind', - best_provider = HuggingFace -) - ### Anthropic ### # claude 3 claude_3_haiku = Model( @@ -704,8 +698,6 @@ class ModelUtils: gemini_1_5_flash.name: gemini_1_5_flash, gemini_2_0_flash.name: gemini_2_0_flash, gemini_2_0_flash_thinking.name: gemini_2_0_flash_thinking, - ### Gemma - gemma_2_27b.name: gemma_2_27b, ### Anthropic ### # claude 3 @@ -799,7 +791,6 @@ class ModelUtils: command_r.name: [command_r, [HuggingSpace]], command_r_plus.name: [command_r_plus, [HuggingSpace]], command_r7b.name: [command_r7b, [HuggingSpace]], - gemma_2_27b.name: [gemma_2_27b, [HuggingFaceAPI]], qwen_2_72b.name: [qwen_2_72b, [HuggingSpace]], qwen_2_5_coder_32b.name: [qwen_2_5_coder_32b, [HuggingFace]], qwq_32b.name: [qwq_32b, [HuggingFace]],