|
|
@ -1,4 +1,4 @@
|
|
|
|
# This file is automatically @generated by Poetry and should not be changed by hand.
|
|
|
|
# This file is automatically @generated by Poetry 1.4.2 and should not be changed by hand.
|
|
|
|
|
|
|
|
|
|
|
|
[[package]]
|
|
|
|
[[package]]
|
|
|
|
name = "absl-py"
|
|
|
|
name = "absl-py"
|
|
|
@ -4005,14 +4005,14 @@ tests = ["pytest", "pytest-mock"]
|
|
|
|
|
|
|
|
|
|
|
|
[[package]]
|
|
|
|
[[package]]
|
|
|
|
name = "langchainplus-sdk"
|
|
|
|
name = "langchainplus-sdk"
|
|
|
|
version = "0.0.6"
|
|
|
|
version = "0.0.7"
|
|
|
|
description = "Client library to connect to the LangChainPlus LLM Tracing and Evaluation Platform."
|
|
|
|
description = "Client library to connect to the LangChainPlus LLM Tracing and Evaluation Platform."
|
|
|
|
category = "main"
|
|
|
|
category = "main"
|
|
|
|
optional = false
|
|
|
|
optional = false
|
|
|
|
python-versions = ">=3.8.1,<4.0"
|
|
|
|
python-versions = ">=3.8.1,<4.0"
|
|
|
|
files = [
|
|
|
|
files = [
|
|
|
|
{file = "langchainplus_sdk-0.0.6-py3-none-any.whl", hash = "sha256:43fe01c66442b88403c969b8812f6be81e023c0d2a6d5d3572a8d87961438658"},
|
|
|
|
{file = "langchainplus_sdk-0.0.7-py3-none-any.whl", hash = "sha256:aefc471058648bf9fc51f659117d33ef905d25a304d5a021f7e32c30f5921076"},
|
|
|
|
{file = "langchainplus_sdk-0.0.6.tar.gz", hash = "sha256:c911a98fd2d02baa48f742b7d700fd6a55f11c9a545ee5d66b08825940c9a32e"},
|
|
|
|
{file = "langchainplus_sdk-0.0.7.tar.gz", hash = "sha256:b58565bdcaf301d2e6e7dd8898f0b8ccf549a35476258e0c14d871d6de02d210"},
|
|
|
|
]
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
|
|
[package.dependencies]
|
|
|
|
[package.dependencies]
|
|
|
@ -11340,13 +11340,13 @@ cffi = {version = ">=1.11", markers = "platform_python_implementation == \"PyPy\
|
|
|
|
cffi = ["cffi (>=1.11)"]
|
|
|
|
cffi = ["cffi (>=1.11)"]
|
|
|
|
|
|
|
|
|
|
|
|
[extras]
|
|
|
|
[extras]
|
|
|
|
all = ["anthropic", "cohere", "openai", "nlpcloud", "huggingface_hub", "jina", "manifest-ml", "elasticsearch", "opensearch-py", "google-search-results", "faiss-cpu", "sentence-transformers", "transformers", "spacy", "nltk", "wikipedia", "beautifulsoup4", "tiktoken", "torch", "jinja2", "pinecone-client", "pinecone-text", "pymongo", "weaviate-client", "redis", "google-api-python-client", "google-auth", "wolframalpha", "qdrant-client", "tensorflow-text", "pypdf", "networkx", "nomic", "aleph-alpha-client", "deeplake", "pgvector", "psycopg2-binary", "pyowm", "pytesseract", "html2text", "atlassian-python-api", "gptcache", "duckduckgo-search", "arxiv", "azure-identity", "clickhouse-connect", "azure-cosmos", "lancedb", "langkit", "lark", "pexpect", "pyvespa", "O365", "jq", "docarray", "steamship", "pdfminer-six", "lxml", "requests-toolbelt", "neo4j", "openlm", "azure-ai-formrecognizer", "azure-ai-vision", "azure-cognitiveservices-speech", "momento", "singlestoredb", "tigrisdb", "nebula3-python"]
|
|
|
|
all = ["O365", "aleph-alpha-client", "anthropic", "arxiv", "atlassian-python-api", "azure-ai-formrecognizer", "azure-ai-vision", "azure-cognitiveservices-speech", "azure-cosmos", "azure-identity", "beautifulsoup4", "clickhouse-connect", "cohere", "deeplake", "docarray", "duckduckgo-search", "elasticsearch", "faiss-cpu", "google-api-python-client", "google-auth", "google-search-results", "gptcache", "html2text", "huggingface_hub", "jina", "jinja2", "jq", "lancedb", "langkit", "lark", "lxml", "manifest-ml", "momento", "nebula3-python", "neo4j", "networkx", "nlpcloud", "nltk", "nomic", "openai", "openlm", "opensearch-py", "pdfminer-six", "pexpect", "pgvector", "pinecone-client", "pinecone-text", "psycopg2-binary", "pymongo", "pyowm", "pypdf", "pytesseract", "pyvespa", "qdrant-client", "redis", "requests-toolbelt", "sentence-transformers", "singlestoredb", "spacy", "steamship", "tensorflow-text", "tigrisdb", "tiktoken", "torch", "transformers", "weaviate-client", "wikipedia", "wolframalpha"]
|
|
|
|
azure = ["azure-identity", "azure-cosmos", "openai", "azure-core", "azure-ai-formrecognizer", "azure-ai-vision", "azure-cognitiveservices-speech"]
|
|
|
|
azure = ["azure-ai-formrecognizer", "azure-ai-vision", "azure-cognitiveservices-speech", "azure-core", "azure-cosmos", "azure-identity", "openai"]
|
|
|
|
cohere = ["cohere"]
|
|
|
|
cohere = ["cohere"]
|
|
|
|
docarray = ["docarray"]
|
|
|
|
docarray = ["docarray"]
|
|
|
|
embeddings = ["sentence-transformers"]
|
|
|
|
embeddings = ["sentence-transformers"]
|
|
|
|
extended-testing = ["beautifulsoup4", "bibtexparser", "chardet", "jq", "pdfminer-six", "pypdf", "pymupdf", "pypdfium2", "tqdm", "lxml", "atlassian-python-api", "beautifulsoup4", "pandas", "telethon", "psychicapi", "zep-python", "gql", "requests-toolbelt", "html2text", "py-trello", "scikit-learn", "pyspark"]
|
|
|
|
extended-testing = ["atlassian-python-api", "beautifulsoup4", "beautifulsoup4", "bibtexparser", "chardet", "gql", "html2text", "jq", "lxml", "pandas", "pdfminer-six", "psychicapi", "py-trello", "pymupdf", "pypdf", "pypdfium2", "pyspark", "requests-toolbelt", "scikit-learn", "telethon", "tqdm", "zep-python"]
|
|
|
|
llms = ["anthropic", "cohere", "openai", "openlm", "nlpcloud", "huggingface_hub", "manifest-ml", "torch", "transformers"]
|
|
|
|
llms = ["anthropic", "cohere", "huggingface_hub", "manifest-ml", "nlpcloud", "openai", "openlm", "torch", "transformers"]
|
|
|
|
openai = ["openai", "tiktoken"]
|
|
|
|
openai = ["openai", "tiktoken"]
|
|
|
|
qdrant = ["qdrant-client"]
|
|
|
|
qdrant = ["qdrant-client"]
|
|
|
|
text-helpers = ["chardet"]
|
|
|
|
text-helpers = ["chardet"]
|
|
|
@ -11354,4 +11354,4 @@ text-helpers = ["chardet"]
|
|
|
|
[metadata]
|
|
|
|
[metadata]
|
|
|
|
lock-version = "2.0"
|
|
|
|
lock-version = "2.0"
|
|
|
|
python-versions = ">=3.8.1,<4.0"
|
|
|
|
python-versions = ">=3.8.1,<4.0"
|
|
|
|
content-hash = "0da3585d7f3216764f396c162c8f9456423b9f80a6dc9af46040c3e5eec0b79e"
|
|
|
|
content-hash = "dbbaa2907bf2ac09ed111ce712772bba0fe56901627f41c53aef71ae5a38d1c6"
|
|
|
|