|
|
|
@ -11,15 +11,18 @@ from .Provider import (
|
|
|
|
|
DeepAi,
|
|
|
|
|
Aivvm,
|
|
|
|
|
Bard,
|
|
|
|
|
H2o
|
|
|
|
|
H2o,
|
|
|
|
|
Yqcloud
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
@dataclass
|
|
|
|
|
class Model:
|
|
|
|
|
name: str
|
|
|
|
|
base_provider: str
|
|
|
|
|
best_provider: Union[type[BaseProvider], tuple[type[BaseProvider]]] = None
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
# Config for HuggingChat, OpenAssistant
|
|
|
|
|
# Works for Liaobots, H2o, OpenaiChat, Yqcloud, You
|
|
|
|
|
default = Model(
|
|
|
|
@ -31,7 +34,7 @@ gpt_35_turbo = Model(
|
|
|
|
|
name='gpt-3.5-turbo',
|
|
|
|
|
base_provider='openai',
|
|
|
|
|
best_provider=[
|
|
|
|
|
DeepAi, CodeLinkAva, ChatgptLogin, ChatgptAi, ChatBase, Aivvm
|
|
|
|
|
DeepAi, CodeLinkAva, ChatgptLogin, ChatgptAi, ChatBase, Aivvm, Yqcloud
|
|
|
|
|
]
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
@ -127,6 +130,12 @@ gpt_35_turbo_16k = Model(
|
|
|
|
|
base_provider='openai',
|
|
|
|
|
best_provider=Vercel)
|
|
|
|
|
|
|
|
|
|
gpt_35_turbo_0613 = Model(
|
|
|
|
|
name='openai:gpt-3.5-turbo-0613',
|
|
|
|
|
base_provider='openai',
|
|
|
|
|
best_provider=[
|
|
|
|
|
Yqcloud, Aivvm, ChatgptLogin])
|
|
|
|
|
|
|
|
|
|
gpt_35_turbo_16k_0613 = Model(
|
|
|
|
|
name='openai:gpt-3.5-turbo-16k-0613',
|
|
|
|
|
base_provider='openai')
|
|
|
|
@ -215,4 +224,5 @@ class ModelUtils:
|
|
|
|
|
'oasst-sft-4-pythia-12b-epoch-3.5': oasst_sft_4_pythia_12b_epoch_35,
|
|
|
|
|
'command-light-nightly': command_light_nightly,
|
|
|
|
|
'gpt-3.5-turbo-16k-0613': gpt_35_turbo_16k_0613,
|
|
|
|
|
'gpt-3.5-turbo-0613': gpt_35_turbo_0613,
|
|
|
|
|
}
|