mirror of
https://github.com/hwchase17/langchain
synced 2024-11-06 03:20:49 +00:00
ed58eeb9c5
Moved the following modules to new package langchain-community in a backwards compatible fashion: ``` mv langchain/langchain/adapters community/langchain_community mv langchain/langchain/callbacks community/langchain_community/callbacks mv langchain/langchain/chat_loaders community/langchain_community mv langchain/langchain/chat_models community/langchain_community mv langchain/langchain/document_loaders community/langchain_community mv langchain/langchain/docstore community/langchain_community mv langchain/langchain/document_transformers community/langchain_community mv langchain/langchain/embeddings community/langchain_community mv langchain/langchain/graphs community/langchain_community mv langchain/langchain/llms community/langchain_community mv langchain/langchain/memory/chat_message_histories community/langchain_community mv langchain/langchain/retrievers community/langchain_community mv langchain/langchain/storage community/langchain_community mv langchain/langchain/tools community/langchain_community mv langchain/langchain/utilities community/langchain_community mv langchain/langchain/vectorstores community/langchain_community mv langchain/langchain/agents/agent_toolkits community/langchain_community mv langchain/langchain/cache.py community/langchain_community mv langchain/langchain/adapters community/langchain_community mv langchain/langchain/callbacks community/langchain_community/callbacks mv langchain/langchain/chat_loaders community/langchain_community mv langchain/langchain/chat_models community/langchain_community mv langchain/langchain/document_loaders community/langchain_community mv langchain/langchain/docstore community/langchain_community mv langchain/langchain/document_transformers community/langchain_community mv langchain/langchain/embeddings community/langchain_community mv langchain/langchain/graphs community/langchain_community mv langchain/langchain/llms community/langchain_community mv langchain/langchain/memory/chat_message_histories community/langchain_community mv langchain/langchain/retrievers community/langchain_community mv langchain/langchain/storage community/langchain_community mv langchain/langchain/tools community/langchain_community mv langchain/langchain/utilities community/langchain_community mv langchain/langchain/vectorstores community/langchain_community mv langchain/langchain/agents/agent_toolkits community/langchain_community mv langchain/langchain/cache.py community/langchain_community ``` Moved the following to core ``` mv langchain/langchain/utils/json_schema.py core/langchain_core/utils mv langchain/langchain/utils/html.py core/langchain_core/utils mv langchain/langchain/utils/strings.py core/langchain_core/utils cat langchain/langchain/utils/env.py >> core/langchain_core/utils/env.py rm langchain/langchain/utils/env.py ``` See .scripts/community_split/script_integrations.sh for all changes
83 lines
2.7 KiB
Python
83 lines
2.7 KiB
Python
"""Test MosaicML API wrapper."""
|
|
import re
|
|
|
|
import pytest
|
|
|
|
from langchain_community.llms.mosaicml import PROMPT_FOR_GENERATION_FORMAT, MosaicML
|
|
|
|
|
|
def test_mosaicml_llm_call() -> None:
|
|
"""Test valid call to MosaicML."""
|
|
llm = MosaicML(model_kwargs={})
|
|
output = llm("Say foo:")
|
|
assert isinstance(output, str)
|
|
|
|
|
|
def test_mosaicml_endpoint_change() -> None:
|
|
"""Test valid call to MosaicML."""
|
|
new_url = "https://models.hosted-on.mosaicml.hosting/mpt-30b-instruct/v1/predict"
|
|
llm = MosaicML(endpoint_url=new_url)
|
|
assert llm.endpoint_url == new_url
|
|
output = llm("Say foo:")
|
|
assert isinstance(output, str)
|
|
|
|
|
|
def test_mosaicml_extra_kwargs() -> None:
|
|
llm = MosaicML(model_kwargs={"max_new_tokens": 1})
|
|
assert llm.model_kwargs == {"max_new_tokens": 1}
|
|
|
|
output = llm("Say foo:")
|
|
|
|
assert isinstance(output, str)
|
|
|
|
# should only generate one new token (which might be a new line or whitespace token)
|
|
assert len(output.split()) <= 1
|
|
|
|
|
|
def test_instruct_prompt() -> None:
|
|
"""Test instruct prompt."""
|
|
llm = MosaicML(inject_instruction_format=True, model_kwargs={"max_new_tokens": 10})
|
|
instruction = "Repeat the word foo"
|
|
prompt = llm._transform_prompt(instruction)
|
|
expected_prompt = PROMPT_FOR_GENERATION_FORMAT.format(instruction=instruction)
|
|
assert prompt == expected_prompt
|
|
output = llm(prompt)
|
|
assert isinstance(output, str)
|
|
|
|
|
|
def test_retry_logic() -> None:
|
|
"""Tests that two queries (which would usually exceed the rate limit) works"""
|
|
llm = MosaicML(inject_instruction_format=True, model_kwargs={"max_new_tokens": 10})
|
|
instruction = "Repeat the word foo"
|
|
prompt = llm._transform_prompt(instruction)
|
|
expected_prompt = PROMPT_FOR_GENERATION_FORMAT.format(instruction=instruction)
|
|
assert prompt == expected_prompt
|
|
output = llm(prompt)
|
|
assert isinstance(output, str)
|
|
output = llm(prompt)
|
|
assert isinstance(output, str)
|
|
|
|
|
|
def test_short_retry_does_not_loop() -> None:
|
|
"""Tests that two queries with a short retry sleep does not infinite loop"""
|
|
llm = MosaicML(
|
|
inject_instruction_format=True,
|
|
model_kwargs={"do_sample": False},
|
|
retry_sleep=0.1,
|
|
)
|
|
instruction = "Repeat the word foo"
|
|
prompt = llm._transform_prompt(instruction)
|
|
expected_prompt = PROMPT_FOR_GENERATION_FORMAT.format(instruction=instruction)
|
|
assert prompt == expected_prompt
|
|
|
|
with pytest.raises(
|
|
ValueError,
|
|
match=re.escape(
|
|
"Error raised by inference API: rate limit exceeded.\nResponse: You have "
|
|
"reached maximum request limit.\n"
|
|
),
|
|
):
|
|
for _ in range(10):
|
|
output = llm(prompt)
|
|
assert isinstance(output, str)
|