mirror of
https://github.com/hwchase17/langchain
synced 2024-11-10 01:10:59 +00:00
fix docs typos (#23668)
Thank you for contributing to LangChain! - [x] **PR title**: "docs: fix multiple typos" Co-authored-by: mohblnk <mohamed.ali@blnk.ai> Co-authored-by: ccurme <chester.curme@gmail.com>
This commit is contained in:
parent
98175860ad
commit
c086410677
@ -207,7 +207,7 @@ class ChatDeepInfra(BaseChatModel):
|
||||
request_timeout: Optional[float] = Field(default=None, alias="timeout")
|
||||
temperature: Optional[float] = 1
|
||||
model_kwargs: Dict[str, Any] = Field(default_factory=dict)
|
||||
"""Run inference with this temperature. Must by in the closed
|
||||
"""Run inference with this temperature. Must be in the closed
|
||||
interval [0.0, 1.0]."""
|
||||
top_p: Optional[float] = None
|
||||
"""Decode using nucleus sampling: consider the smallest set of tokens whose
|
||||
|
@ -236,7 +236,7 @@ class ChatGooglePalm(BaseChatModel, BaseModel):
|
||||
"""Model name to use."""
|
||||
google_api_key: Optional[SecretStr] = None
|
||||
temperature: Optional[float] = None
|
||||
"""Run inference with this temperature. Must by in the closed
|
||||
"""Run inference with this temperature. Must be in the closed
|
||||
interval [0.0, 1.0]."""
|
||||
top_p: Optional[float] = None
|
||||
"""Decode using nucleus sampling: consider the smallest set of tokens whose
|
||||
|
@ -190,7 +190,7 @@ class ChatLiteLLM(BaseChatModel):
|
||||
request_timeout: Optional[Union[float, Tuple[float, float]]] = None
|
||||
temperature: Optional[float] = 1
|
||||
model_kwargs: Dict[str, Any] = Field(default_factory=dict)
|
||||
"""Run inference with this temperature. Must by in the closed
|
||||
"""Run inference with this temperature. Must be in the closed
|
||||
interval [0.0, 1.0]."""
|
||||
top_p: Optional[float] = None
|
||||
"""Decode using nucleus sampling: consider the smallest set of tokens whose
|
||||
|
@ -72,7 +72,7 @@ class GooglePalm(BaseLLM, BaseModel):
|
||||
model_name: str = "models/text-bison-001"
|
||||
"""Model name to use."""
|
||||
temperature: float = 0.7
|
||||
"""Run inference with this temperature. Must by in the closed interval
|
||||
"""Run inference with this temperature. Must be in the closed interval
|
||||
[0.0, 1.0]."""
|
||||
top_p: Optional[float] = None
|
||||
"""Decode using nucleus sampling: consider the smallest set of tokens whose
|
||||
|
Loading…
Reference in New Issue
Block a user