From 2befff83d68bb031e3c6a79fb4a37a8e17614d76 Mon Sep 17 00:00:00 2001 From: Andriy Mulyar Date: Mon, 24 Jul 2023 12:01:37 -0400 Subject: [PATCH] top_p error in gpt4all-api --- gpt4all-api/gpt4all_api/app/api_v1/routes/completions.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/gpt4all-api/gpt4all_api/app/api_v1/routes/completions.py b/gpt4all-api/gpt4all_api/app/api_v1/routes/completions.py index 5df9ddb7..31ed82ed 100644 --- a/gpt4all-api/gpt4all_api/app/api_v1/routes/completions.py +++ b/gpt4all-api/gpt4all_api/app/api_v1/routes/completions.py @@ -24,7 +24,7 @@ class CompletionRequest(BaseModel): prompt: Union[List[str], str] = Field(..., description='The prompt to begin completing from.') max_tokens: int = Field(None, description='Max tokens to generate') temperature: float = Field(settings.temp, description='Model temperature') - top_p: float = Field(settings.top_k, description='top_p') + top_p: float = Field(settings.top_p, description='top_p') top_k: int = Field(settings.top_k, description='top_k') n: int = Field(1, description='How many completions to generate for each prompt') stream: bool = Field(False, description='Stream responses')