diff --git a/g4f/Provider/HuggingChat.py b/g4f/Provider/HuggingChat.py index 8598358d..06216ade 100644 --- a/g4f/Provider/HuggingChat.py +++ b/g4f/Provider/HuggingChat.py @@ -26,7 +26,7 @@ class HuggingChat(AbstractProvider, ProviderModelMixin): "llama-3.1-70b": "meta-llama/Meta-Llama-3.1-70B-Instruct", "command-r-plus": "CohereForAI/c4ai-command-r-plus-08-2024", "mixtral-8x7b": "mistralai/Mixtral-8x7B-Instruct-v0.1", - "mixtral-8x7b": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO", + "mixtral-8x7b-dpo": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO", "mistral-7b": "mistralai/Mistral-7B-Instruct-v0.3", "phi-3-mini-4k": "microsoft/Phi-3-mini-4k-instruct", } diff --git a/g4f/models.py b/g4f/models.py index 0c3c8402..76aba557 100644 --- a/g4f/models.py +++ b/g4f/models.py @@ -189,6 +189,7 @@ llama_3_1_405b = Model( best_provider = IterListProvider([Blackbox]) ) + ### Mistral ### mixtral_8x7b = Model( name = "mixtral-8x7b", @@ -203,6 +204,14 @@ mistral_7b = Model( ) +### NousResearch ### +mixtral_8x7b_dpo = Model( + name = "mixtral-8x7b-dpo", + base_provider = "NousResearch", + best_provider = IterListProvider([HuggingChat, HuggingFace,]) +) + + ### Microsoft ### phi_3_mini_4k = Model( name = "phi-3-mini-4k",