2023-04-30 18:14:09 +00:00
|
|
|
"""Integration tests for the langchain tracer module."""
|
|
|
|
import asyncio
|
|
|
|
import os
|
|
|
|
|
|
|
|
from aiohttp import ClientSession
|
2023-12-11 21:53:30 +00:00
|
|
|
from langchain_core.callbacks.manager import atrace_as_chain_group, trace_as_chain_group
|
2023-11-20 21:09:30 +00:00
|
|
|
from langchain_core.prompts import PromptTemplate
|
2023-12-11 21:53:30 +00:00
|
|
|
from langchain_core.tracers.context import tracing_enabled, tracing_v2_enabled
|
2023-04-30 18:14:09 +00:00
|
|
|
|
2023-12-11 21:53:30 +00:00
|
|
|
from langchain_community.chat_models import ChatOpenAI
|
|
|
|
from langchain_community.llms import OpenAI
|
2023-04-30 18:14:09 +00:00
|
|
|
|
|
|
|
questions = [
|
|
|
|
(
|
|
|
|
"Who won the US Open men's final in 2019? "
|
|
|
|
"What is his age raised to the 0.334 power?"
|
|
|
|
),
|
|
|
|
(
|
|
|
|
"Who is Olivia Wilde's boyfriend? "
|
|
|
|
"What is his current age raised to the 0.23 power?"
|
|
|
|
),
|
|
|
|
(
|
|
|
|
"Who won the most recent formula 1 grand prix? "
|
|
|
|
"What is their age raised to the 0.23 power?"
|
|
|
|
),
|
|
|
|
(
|
|
|
|
"Who won the US Open women's final in 2019? "
|
|
|
|
"What is her age raised to the 0.34 power?"
|
|
|
|
),
|
|
|
|
("Who is Beyonce's husband? " "What is his age raised to the 0.19 power?"),
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
|
|
def test_tracing_sequential() -> None:
|
2023-12-11 21:53:30 +00:00
|
|
|
from langchain.agents import AgentType, initialize_agent, load_tools
|
|
|
|
|
2023-04-30 18:14:09 +00:00
|
|
|
os.environ["LANGCHAIN_TRACING"] = "true"
|
|
|
|
|
|
|
|
for q in questions[:3]:
|
|
|
|
llm = OpenAI(temperature=0)
|
|
|
|
tools = load_tools(["llm-math", "serpapi"], llm=llm)
|
|
|
|
agent = initialize_agent(
|
|
|
|
tools, llm, agent=AgentType.ZERO_SHOT_REACT_DESCRIPTION, verbose=True
|
|
|
|
)
|
|
|
|
agent.run(q)
|
|
|
|
|
|
|
|
|
|
|
|
def test_tracing_session_env_var() -> None:
|
2023-12-11 21:53:30 +00:00
|
|
|
from langchain.agents import AgentType, initialize_agent, load_tools
|
|
|
|
|
2023-04-30 18:14:09 +00:00
|
|
|
os.environ["LANGCHAIN_TRACING"] = "true"
|
|
|
|
os.environ["LANGCHAIN_SESSION"] = "my_session"
|
|
|
|
|
|
|
|
llm = OpenAI(temperature=0)
|
|
|
|
tools = load_tools(["llm-math", "serpapi"], llm=llm)
|
|
|
|
agent = initialize_agent(
|
|
|
|
tools, llm, agent=AgentType.ZERO_SHOT_REACT_DESCRIPTION, verbose=True
|
|
|
|
)
|
|
|
|
agent.run(questions[0])
|
|
|
|
if "LANGCHAIN_SESSION" in os.environ:
|
|
|
|
del os.environ["LANGCHAIN_SESSION"]
|
|
|
|
|
|
|
|
|
|
|
|
async def test_tracing_concurrent() -> None:
|
2023-12-11 21:53:30 +00:00
|
|
|
from langchain.agents import AgentType, initialize_agent, load_tools
|
|
|
|
|
2023-04-30 18:14:09 +00:00
|
|
|
os.environ["LANGCHAIN_TRACING"] = "true"
|
|
|
|
aiosession = ClientSession()
|
|
|
|
llm = OpenAI(temperature=0)
|
|
|
|
async_tools = load_tools(["llm-math", "serpapi"], llm=llm, aiosession=aiosession)
|
|
|
|
agent = initialize_agent(
|
|
|
|
async_tools, llm, agent=AgentType.ZERO_SHOT_REACT_DESCRIPTION, verbose=True
|
|
|
|
)
|
|
|
|
tasks = [agent.arun(q) for q in questions[:3]]
|
|
|
|
await asyncio.gather(*tasks)
|
|
|
|
await aiosession.close()
|
|
|
|
|
|
|
|
|
|
|
|
async def test_tracing_concurrent_bw_compat_environ() -> None:
|
2023-12-11 21:53:30 +00:00
|
|
|
from langchain.agents import AgentType, initialize_agent, load_tools
|
|
|
|
|
2023-04-30 18:14:09 +00:00
|
|
|
os.environ["LANGCHAIN_HANDLER"] = "langchain"
|
|
|
|
if "LANGCHAIN_TRACING" in os.environ:
|
|
|
|
del os.environ["LANGCHAIN_TRACING"]
|
|
|
|
aiosession = ClientSession()
|
|
|
|
llm = OpenAI(temperature=0)
|
|
|
|
async_tools = load_tools(["llm-math", "serpapi"], llm=llm, aiosession=aiosession)
|
|
|
|
agent = initialize_agent(
|
|
|
|
async_tools, llm, agent=AgentType.ZERO_SHOT_REACT_DESCRIPTION, verbose=True
|
|
|
|
)
|
|
|
|
tasks = [agent.arun(q) for q in questions[:3]]
|
|
|
|
await asyncio.gather(*tasks)
|
|
|
|
await aiosession.close()
|
|
|
|
if "LANGCHAIN_HANDLER" in os.environ:
|
|
|
|
del os.environ["LANGCHAIN_HANDLER"]
|
|
|
|
|
|
|
|
|
|
|
|
def test_tracing_context_manager() -> None:
|
2023-12-11 21:53:30 +00:00
|
|
|
from langchain.agents import AgentType, initialize_agent, load_tools
|
|
|
|
|
2023-04-30 18:14:09 +00:00
|
|
|
llm = OpenAI(temperature=0)
|
|
|
|
tools = load_tools(["llm-math", "serpapi"], llm=llm)
|
|
|
|
agent = initialize_agent(
|
|
|
|
tools, llm, agent=AgentType.ZERO_SHOT_REACT_DESCRIPTION, verbose=True
|
|
|
|
)
|
|
|
|
if "LANGCHAIN_TRACING" in os.environ:
|
|
|
|
del os.environ["LANGCHAIN_TRACING"]
|
|
|
|
with tracing_enabled() as session:
|
|
|
|
assert session
|
|
|
|
agent.run(questions[0]) # this should be traced
|
|
|
|
|
|
|
|
agent.run(questions[0]) # this should not be traced
|
|
|
|
|
|
|
|
|
|
|
|
async def test_tracing_context_manager_async() -> None:
|
2023-12-11 21:53:30 +00:00
|
|
|
from langchain.agents import AgentType, initialize_agent, load_tools
|
|
|
|
|
2023-04-30 18:14:09 +00:00
|
|
|
llm = OpenAI(temperature=0)
|
|
|
|
async_tools = load_tools(["llm-math", "serpapi"], llm=llm)
|
|
|
|
agent = initialize_agent(
|
|
|
|
async_tools, llm, agent=AgentType.ZERO_SHOT_REACT_DESCRIPTION, verbose=True
|
|
|
|
)
|
|
|
|
if "LANGCHAIN_TRACING" in os.environ:
|
|
|
|
del os.environ["LANGCHAIN_TRACING"]
|
|
|
|
|
|
|
|
# start a background task
|
|
|
|
task = asyncio.create_task(agent.arun(questions[0])) # this should not be traced
|
|
|
|
with tracing_enabled() as session:
|
|
|
|
assert session
|
|
|
|
tasks = [agent.arun(q) for q in questions[1:4]] # these should be traced
|
|
|
|
await asyncio.gather(*tasks)
|
|
|
|
|
|
|
|
await task
|
2023-05-10 18:08:29 +00:00
|
|
|
|
|
|
|
|
|
|
|
async def test_tracing_v2_environment_variable() -> None:
|
2023-12-11 21:53:30 +00:00
|
|
|
from langchain.agents import AgentType, initialize_agent, load_tools
|
|
|
|
|
2023-05-10 18:08:29 +00:00
|
|
|
os.environ["LANGCHAIN_TRACING_V2"] = "true"
|
|
|
|
|
|
|
|
aiosession = ClientSession()
|
|
|
|
llm = OpenAI(temperature=0)
|
|
|
|
async_tools = load_tools(["llm-math", "serpapi"], llm=llm, aiosession=aiosession)
|
|
|
|
agent = initialize_agent(
|
|
|
|
async_tools, llm, agent=AgentType.ZERO_SHOT_REACT_DESCRIPTION, verbose=True
|
|
|
|
)
|
|
|
|
tasks = [agent.arun(q) for q in questions[:3]]
|
|
|
|
await asyncio.gather(*tasks)
|
|
|
|
await aiosession.close()
|
|
|
|
|
|
|
|
|
|
|
|
def test_tracing_v2_context_manager() -> None:
|
2023-12-11 21:53:30 +00:00
|
|
|
from langchain.agents import AgentType, initialize_agent, load_tools
|
|
|
|
|
2023-05-31 01:47:06 +00:00
|
|
|
llm = ChatOpenAI(temperature=0)
|
2023-05-10 18:08:29 +00:00
|
|
|
tools = load_tools(["llm-math", "serpapi"], llm=llm)
|
|
|
|
agent = initialize_agent(
|
2023-05-31 01:47:06 +00:00
|
|
|
tools, llm, agent=AgentType.CHAT_ZERO_SHOT_REACT_DESCRIPTION, verbose=True
|
2023-05-10 18:08:29 +00:00
|
|
|
)
|
|
|
|
if "LANGCHAIN_TRACING_V2" in os.environ:
|
|
|
|
del os.environ["LANGCHAIN_TRACING_V2"]
|
2023-06-02 19:13:56 +00:00
|
|
|
with tracing_v2_enabled():
|
2023-05-10 18:08:29 +00:00
|
|
|
agent.run(questions[0]) # this should be traced
|
|
|
|
|
|
|
|
agent.run(questions[0]) # this should not be traced
|
2023-06-06 02:18:43 +00:00
|
|
|
|
|
|
|
|
Add support for tags (#5898)
<!--
Thank you for contributing to LangChain! Your PR will appear in our
release under the title you set. Please make sure it highlights your
valuable contribution.
Replace this with a description of the change, the issue it fixes (if
applicable), and relevant context. List any dependencies required for
this change.
After you're done, someone will review your PR. They may suggest
improvements. If no one reviews your PR within a few days, feel free to
@-mention the same people again, as notifications can get lost.
Finally, we'd love to show appreciation for your contribution - if you'd
like us to shout you out on Twitter, please also include your handle!
-->
<!-- Remove if not applicable -->
Fixes # (issue)
#### Before submitting
<!-- If you're adding a new integration, please include:
1. a test for the integration - favor unit tests that does not rely on
network access.
2. an example notebook showing its use
See contribution guidelines for more information on how to write tests,
lint
etc:
https://github.com/hwchase17/langchain/blob/master/.github/CONTRIBUTING.md
-->
#### Who can review?
Tag maintainers/contributors who might be interested:
<!-- For a quicker response, figure out the right person to tag with @
@hwchase17 - project lead
Tracing / Callbacks
- @agola11
Async
- @agola11
DataLoaders
- @eyurtsev
Models
- @hwchase17
- @agola11
Agents / Tools / Toolkits
- @vowelparrot
VectorStores / Retrievers / Memory
- @dev2049
-->
2023-06-13 19:30:59 +00:00
|
|
|
def test_tracing_v2_chain_with_tags() -> None:
|
2023-12-11 21:53:30 +00:00
|
|
|
from langchain.chains.constitutional_ai.base import ConstitutionalChain
|
|
|
|
from langchain.chains.constitutional_ai.models import ConstitutionalPrinciple
|
|
|
|
from langchain.chains.llm import LLMChain
|
|
|
|
|
Add support for tags (#5898)
<!--
Thank you for contributing to LangChain! Your PR will appear in our
release under the title you set. Please make sure it highlights your
valuable contribution.
Replace this with a description of the change, the issue it fixes (if
applicable), and relevant context. List any dependencies required for
this change.
After you're done, someone will review your PR. They may suggest
improvements. If no one reviews your PR within a few days, feel free to
@-mention the same people again, as notifications can get lost.
Finally, we'd love to show appreciation for your contribution - if you'd
like us to shout you out on Twitter, please also include your handle!
-->
<!-- Remove if not applicable -->
Fixes # (issue)
#### Before submitting
<!-- If you're adding a new integration, please include:
1. a test for the integration - favor unit tests that does not rely on
network access.
2. an example notebook showing its use
See contribution guidelines for more information on how to write tests,
lint
etc:
https://github.com/hwchase17/langchain/blob/master/.github/CONTRIBUTING.md
-->
#### Who can review?
Tag maintainers/contributors who might be interested:
<!-- For a quicker response, figure out the right person to tag with @
@hwchase17 - project lead
Tracing / Callbacks
- @agola11
Async
- @agola11
DataLoaders
- @eyurtsev
Models
- @hwchase17
- @agola11
Agents / Tools / Toolkits
- @vowelparrot
VectorStores / Retrievers / Memory
- @dev2049
-->
2023-06-13 19:30:59 +00:00
|
|
|
llm = OpenAI(temperature=0)
|
|
|
|
chain = ConstitutionalChain.from_llm(
|
|
|
|
llm,
|
|
|
|
chain=LLMChain.from_string(llm, "Q: {question} A:"),
|
|
|
|
tags=["only-root"],
|
|
|
|
constitutional_principles=[
|
|
|
|
ConstitutionalPrinciple(
|
|
|
|
critique_request="Tell if this answer is good.",
|
|
|
|
revision_request="Give a better answer.",
|
|
|
|
)
|
|
|
|
],
|
|
|
|
)
|
|
|
|
if "LANGCHAIN_TRACING_V2" in os.environ:
|
|
|
|
del os.environ["LANGCHAIN_TRACING_V2"]
|
|
|
|
with tracing_v2_enabled():
|
|
|
|
chain.run("what is the meaning of life", tags=["a-tag"])
|
|
|
|
|
|
|
|
|
support adding custom metadata to runs (#7120)
- [x] wire up tools
- [x] wire up retrievers
- [x] add integration test
<!-- Thank you for contributing to LangChain!
Replace this comment with:
- Description: a description of the change,
- Issue: the issue # it fixes (if applicable),
- Dependencies: any dependencies required for this change,
- Tag maintainer: for a quicker response, tag the relevant maintainer
(see below),
- Twitter handle: we announce bigger features on Twitter. If your PR
gets announced and you'd like a mention, we'll gladly shout you out!
If you're adding a new integration, please include:
1. a test for the integration, preferably unit tests that do not rely on
network access,
2. an example notebook showing its use.
Maintainer responsibilities:
- General / Misc / if you don't know who to tag: @baskaryan
- DataLoaders / VectorStores / Retrievers: @rlancemartin, @eyurtsev
- Models / Prompts: @hwchase17, @baskaryan
- Memory: @hwchase17
- Agents / Tools / Toolkits: @hinthornw
- Tracing / Callbacks: @agola11
- Async: @agola11
If no one reviews your PR within a few days, feel free to @-mention the
same people again.
See contribution guidelines for more information on how to write/run
tests, lint, etc:
https://github.com/hwchase17/langchain/blob/master/.github/CONTRIBUTING.md
-->
2023-07-05 18:11:38 +00:00
|
|
|
def test_tracing_v2_agent_with_metadata() -> None:
|
2023-12-11 21:53:30 +00:00
|
|
|
from langchain.agents import AgentType, initialize_agent, load_tools
|
|
|
|
|
support adding custom metadata to runs (#7120)
- [x] wire up tools
- [x] wire up retrievers
- [x] add integration test
<!-- Thank you for contributing to LangChain!
Replace this comment with:
- Description: a description of the change,
- Issue: the issue # it fixes (if applicable),
- Dependencies: any dependencies required for this change,
- Tag maintainer: for a quicker response, tag the relevant maintainer
(see below),
- Twitter handle: we announce bigger features on Twitter. If your PR
gets announced and you'd like a mention, we'll gladly shout you out!
If you're adding a new integration, please include:
1. a test for the integration, preferably unit tests that do not rely on
network access,
2. an example notebook showing its use.
Maintainer responsibilities:
- General / Misc / if you don't know who to tag: @baskaryan
- DataLoaders / VectorStores / Retrievers: @rlancemartin, @eyurtsev
- Models / Prompts: @hwchase17, @baskaryan
- Memory: @hwchase17
- Agents / Tools / Toolkits: @hinthornw
- Tracing / Callbacks: @agola11
- Async: @agola11
If no one reviews your PR within a few days, feel free to @-mention the
same people again.
See contribution guidelines for more information on how to write/run
tests, lint, etc:
https://github.com/hwchase17/langchain/blob/master/.github/CONTRIBUTING.md
-->
2023-07-05 18:11:38 +00:00
|
|
|
os.environ["LANGCHAIN_TRACING_V2"] = "true"
|
|
|
|
llm = OpenAI(temperature=0)
|
|
|
|
chat = ChatOpenAI(temperature=0)
|
|
|
|
tools = load_tools(["llm-math", "serpapi"], llm=llm)
|
|
|
|
agent = initialize_agent(
|
|
|
|
tools, llm, agent=AgentType.ZERO_SHOT_REACT_DESCRIPTION, verbose=True
|
|
|
|
)
|
|
|
|
chat_agent = initialize_agent(
|
|
|
|
tools, chat, agent=AgentType.CHAT_ZERO_SHOT_REACT_DESCRIPTION, verbose=True
|
|
|
|
)
|
|
|
|
agent.run(questions[0], tags=["a-tag"], metadata={"a": "b", "c": "d"})
|
|
|
|
chat_agent.run(questions[0], tags=["a-tag"], metadata={"a": "b", "c": "d"})
|
|
|
|
|
|
|
|
|
|
|
|
async def test_tracing_v2_async_agent_with_metadata() -> None:
|
2023-12-11 21:53:30 +00:00
|
|
|
from langchain.agents import AgentType, initialize_agent, load_tools
|
|
|
|
|
support adding custom metadata to runs (#7120)
- [x] wire up tools
- [x] wire up retrievers
- [x] add integration test
<!-- Thank you for contributing to LangChain!
Replace this comment with:
- Description: a description of the change,
- Issue: the issue # it fixes (if applicable),
- Dependencies: any dependencies required for this change,
- Tag maintainer: for a quicker response, tag the relevant maintainer
(see below),
- Twitter handle: we announce bigger features on Twitter. If your PR
gets announced and you'd like a mention, we'll gladly shout you out!
If you're adding a new integration, please include:
1. a test for the integration, preferably unit tests that do not rely on
network access,
2. an example notebook showing its use.
Maintainer responsibilities:
- General / Misc / if you don't know who to tag: @baskaryan
- DataLoaders / VectorStores / Retrievers: @rlancemartin, @eyurtsev
- Models / Prompts: @hwchase17, @baskaryan
- Memory: @hwchase17
- Agents / Tools / Toolkits: @hinthornw
- Tracing / Callbacks: @agola11
- Async: @agola11
If no one reviews your PR within a few days, feel free to @-mention the
same people again.
See contribution guidelines for more information on how to write/run
tests, lint, etc:
https://github.com/hwchase17/langchain/blob/master/.github/CONTRIBUTING.md
-->
2023-07-05 18:11:38 +00:00
|
|
|
os.environ["LANGCHAIN_TRACING_V2"] = "true"
|
|
|
|
llm = OpenAI(temperature=0, metadata={"f": "g", "h": "i"})
|
|
|
|
chat = ChatOpenAI(temperature=0, metadata={"f": "g", "h": "i"})
|
|
|
|
async_tools = load_tools(["llm-math", "serpapi"], llm=llm)
|
|
|
|
agent = initialize_agent(
|
|
|
|
async_tools, llm, agent=AgentType.ZERO_SHOT_REACT_DESCRIPTION, verbose=True
|
|
|
|
)
|
|
|
|
chat_agent = initialize_agent(
|
|
|
|
async_tools,
|
|
|
|
chat,
|
|
|
|
agent=AgentType.CHAT_ZERO_SHOT_REACT_DESCRIPTION,
|
|
|
|
verbose=True,
|
|
|
|
)
|
|
|
|
await agent.arun(questions[0], tags=["a-tag"], metadata={"a": "b", "c": "d"})
|
|
|
|
await chat_agent.arun(questions[0], tags=["a-tag"], metadata={"a": "b", "c": "d"})
|
|
|
|
|
|
|
|
|
2023-06-06 02:18:43 +00:00
|
|
|
def test_trace_as_group() -> None:
|
2023-12-11 21:53:30 +00:00
|
|
|
from langchain.chains.llm import LLMChain
|
|
|
|
|
2023-06-06 02:18:43 +00:00
|
|
|
llm = OpenAI(temperature=0.9)
|
|
|
|
prompt = PromptTemplate(
|
|
|
|
input_variables=["product"],
|
|
|
|
template="What is a good name for a company that makes {product}?",
|
|
|
|
)
|
|
|
|
chain = LLMChain(llm=llm, prompt=prompt)
|
2023-09-17 07:47:51 +00:00
|
|
|
with trace_as_chain_group("my_group", inputs={"input": "cars"}) as group_manager:
|
2023-06-06 02:18:43 +00:00
|
|
|
chain.run(product="cars", callbacks=group_manager)
|
|
|
|
chain.run(product="computers", callbacks=group_manager)
|
2023-09-17 07:47:51 +00:00
|
|
|
final_res = chain.run(product="toys", callbacks=group_manager)
|
|
|
|
group_manager.on_chain_end({"output": final_res})
|
2023-06-06 02:18:43 +00:00
|
|
|
|
2023-09-17 07:47:51 +00:00
|
|
|
with trace_as_chain_group("my_group_2", inputs={"input": "toys"}) as group_manager:
|
|
|
|
final_res = chain.run(product="toys", callbacks=group_manager)
|
|
|
|
group_manager.on_chain_end({"output": final_res})
|
2023-06-06 02:18:43 +00:00
|
|
|
|
|
|
|
|
|
|
|
def test_trace_as_group_with_env_set() -> None:
|
2023-12-11 21:53:30 +00:00
|
|
|
from langchain.chains.llm import LLMChain
|
|
|
|
|
2023-06-06 02:18:43 +00:00
|
|
|
os.environ["LANGCHAIN_TRACING_V2"] = "true"
|
|
|
|
llm = OpenAI(temperature=0.9)
|
|
|
|
prompt = PromptTemplate(
|
|
|
|
input_variables=["product"],
|
|
|
|
template="What is a good name for a company that makes {product}?",
|
|
|
|
)
|
|
|
|
chain = LLMChain(llm=llm, prompt=prompt)
|
2023-09-17 07:47:51 +00:00
|
|
|
with trace_as_chain_group(
|
|
|
|
"my_group_env_set", inputs={"input": "cars"}
|
|
|
|
) as group_manager:
|
2023-06-06 02:18:43 +00:00
|
|
|
chain.run(product="cars", callbacks=group_manager)
|
|
|
|
chain.run(product="computers", callbacks=group_manager)
|
2023-09-17 07:47:51 +00:00
|
|
|
final_res = chain.run(product="toys", callbacks=group_manager)
|
|
|
|
group_manager.on_chain_end({"output": final_res})
|
2023-06-06 02:18:43 +00:00
|
|
|
|
2023-09-17 07:47:51 +00:00
|
|
|
with trace_as_chain_group(
|
|
|
|
"my_group_2_env_set", inputs={"input": "toys"}
|
|
|
|
) as group_manager:
|
|
|
|
final_res = chain.run(product="toys", callbacks=group_manager)
|
|
|
|
group_manager.on_chain_end({"output": final_res})
|
2023-06-06 02:18:43 +00:00
|
|
|
|
|
|
|
|
|
|
|
async def test_trace_as_group_async() -> None:
|
2023-12-11 21:53:30 +00:00
|
|
|
from langchain.chains.llm import LLMChain
|
|
|
|
|
2023-06-06 02:18:43 +00:00
|
|
|
llm = OpenAI(temperature=0.9)
|
|
|
|
prompt = PromptTemplate(
|
|
|
|
input_variables=["product"],
|
|
|
|
template="What is a good name for a company that makes {product}?",
|
|
|
|
)
|
|
|
|
chain = LLMChain(llm=llm, prompt=prompt)
|
2023-09-17 07:47:51 +00:00
|
|
|
async with atrace_as_chain_group("my_async_group") as group_manager:
|
2023-06-06 02:18:43 +00:00
|
|
|
await chain.arun(product="cars", callbacks=group_manager)
|
|
|
|
await chain.arun(product="computers", callbacks=group_manager)
|
|
|
|
await chain.arun(product="toys", callbacks=group_manager)
|
|
|
|
|
2023-09-17 07:47:51 +00:00
|
|
|
async with atrace_as_chain_group(
|
|
|
|
"my_async_group_2", inputs={"input": "toys"}
|
|
|
|
) as group_manager:
|
|
|
|
res = await asyncio.gather(
|
2023-06-06 02:18:43 +00:00
|
|
|
*[
|
|
|
|
chain.arun(product="toys", callbacks=group_manager),
|
|
|
|
chain.arun(product="computers", callbacks=group_manager),
|
|
|
|
chain.arun(product="cars", callbacks=group_manager),
|
|
|
|
]
|
|
|
|
)
|
2023-09-17 07:47:51 +00:00
|
|
|
await group_manager.on_chain_end({"output": res})
|