mirror of
https://github.com/hwchase17/langchain
synced 2024-11-10 01:10:59 +00:00
98 lines
3.3 KiB
Python
98 lines
3.3 KiB
Python
|
"""Test Friendli chat API."""
|
||
|
|
||
|
import pytest
|
||
|
from langchain_core.messages.ai import AIMessage
|
||
|
from langchain_core.messages.human import HumanMessage
|
||
|
from langchain_core.outputs.generation import Generation
|
||
|
from langchain_core.outputs.llm_result import LLMResult
|
||
|
|
||
|
from langchain_community.chat_models.friendli import ChatFriendli
|
||
|
|
||
|
|
||
|
@pytest.fixture
|
||
|
def friendli_chat() -> ChatFriendli:
|
||
|
"""Friendli LLM for chat."""
|
||
|
return ChatFriendli(temperature=0, max_tokens=10)
|
||
|
|
||
|
|
||
|
def test_friendli_invoke(friendli_chat: ChatFriendli) -> None:
|
||
|
"""Test invoke."""
|
||
|
output = friendli_chat.invoke("What is generative AI?")
|
||
|
assert isinstance(output, AIMessage)
|
||
|
assert isinstance(output.content, str)
|
||
|
|
||
|
|
||
|
async def test_friendli_ainvoke(friendli_chat: ChatFriendli) -> None:
|
||
|
"""Test async invoke."""
|
||
|
output = await friendli_chat.ainvoke("What is generative AI?")
|
||
|
assert isinstance(output, AIMessage)
|
||
|
assert isinstance(output.content, str)
|
||
|
|
||
|
|
||
|
def test_friendli_batch(friendli_chat: ChatFriendli) -> None:
|
||
|
"""Test batch."""
|
||
|
outputs = friendli_chat.batch(["What is generative AI?", "What is generative AI?"])
|
||
|
for output in outputs:
|
||
|
assert isinstance(output, AIMessage)
|
||
|
assert isinstance(output.content, str)
|
||
|
|
||
|
|
||
|
async def test_friendli_abatch(friendli_chat: ChatFriendli) -> None:
|
||
|
"""Test async batch."""
|
||
|
outputs = await friendli_chat.abatch(
|
||
|
["What is generative AI?", "What is generative AI?"]
|
||
|
)
|
||
|
for output in outputs:
|
||
|
assert isinstance(output, AIMessage)
|
||
|
assert isinstance(output.content, str)
|
||
|
|
||
|
|
||
|
def test_friendli_generate(friendli_chat: ChatFriendli) -> None:
|
||
|
"""Test generate."""
|
||
|
message = HumanMessage(content="What is generative AI?")
|
||
|
result = friendli_chat.generate([[message], [message]])
|
||
|
assert isinstance(result, LLMResult)
|
||
|
generations = result.generations
|
||
|
assert len(generations) == 2
|
||
|
for generation in generations:
|
||
|
gen_ = generation[0]
|
||
|
assert isinstance(gen_, Generation)
|
||
|
text = gen_.text
|
||
|
assert isinstance(text, str)
|
||
|
generation_info = gen_.generation_info
|
||
|
if generation_info is not None:
|
||
|
assert "token" in generation_info
|
||
|
|
||
|
|
||
|
async def test_friendli_agenerate(friendli_chat: ChatFriendli) -> None:
|
||
|
"""Test async generate."""
|
||
|
message = HumanMessage(content="What is generative AI?")
|
||
|
result = await friendli_chat.agenerate([[message], [message]])
|
||
|
assert isinstance(result, LLMResult)
|
||
|
generations = result.generations
|
||
|
assert len(generations) == 2
|
||
|
for generation in generations:
|
||
|
gen_ = generation[0]
|
||
|
assert isinstance(gen_, Generation)
|
||
|
text = gen_.text
|
||
|
assert isinstance(text, str)
|
||
|
generation_info = gen_.generation_info
|
||
|
if generation_info is not None:
|
||
|
assert "token" in generation_info
|
||
|
|
||
|
|
||
|
def test_friendli_stream(friendli_chat: ChatFriendli) -> None:
|
||
|
"""Test stream."""
|
||
|
stream = friendli_chat.stream("Say hello world.")
|
||
|
for chunk in stream:
|
||
|
assert isinstance(chunk, AIMessage)
|
||
|
assert isinstance(chunk.content, str)
|
||
|
|
||
|
|
||
|
async def test_friendli_astream(friendli_chat: ChatFriendli) -> None:
|
||
|
"""Test async stream."""
|
||
|
stream = friendli_chat.astream("Say hello world.")
|
||
|
async for chunk in stream:
|
||
|
assert isinstance(chunk, AIMessage)
|
||
|
assert isinstance(chunk.content, str)
|