community[minor]: Add OCI Generative AI integration (#16548)
<!-- Thank you for contributing to LangChain!
Please title your PR "<package>: <description>", where <package> is
whichever of langchain, community, core, experimental, etc. is being
modified.
Replace this entire comment with:
- **Description:** Adding Oracle Cloud Infrastructure Generative AI
integration. Oracle Cloud Infrastructure (OCI) Generative AI is a fully
managed service that provides a set of state-of-the-art, customizable
large language models (LLMs) that cover a wide range of use cases, and
which is available through a single API. Using the OCI Generative AI
service you can access ready-to-use pretrained models, or create and
host your own fine-tuned custom models based on your own data on
dedicated AI clusters.
https://docs.oracle.com/en-us/iaas/Content/generative-ai/home.htm
- **Issue:** None,
- **Dependencies:** OCI Python SDK,
- **Twitter handle:** we announce bigger features on Twitter. If your PR
gets announced, and you'd like a mention, we'll gladly shout you out!
Please make sure your PR is passing linting and testing before
submitting. Run `make format`, `make lint` and `make test` from the root
of the package you've modified to check this locally.
Passed
See contribution guidelines for more information on how to write/run
tests, lint, etc: https://python.langchain.com/docs/contributing/
If you're adding a new integration, please include:
1. a test for the integration, preferably unit tests that do not rely on
network access,
2. an example notebook showing its use. It lives in
`docs/docs/integrations` directory.
we provide unit tests. However, we cannot provide integration tests due
to Oracle policies that prohibit public sharing of api keys.
If no one reviews your PR within a few days, please @-mention one of
@baskaryan, @eyurtsev, @hwchase17.
-->
---------
Co-authored-by: Arthur Cheng <arthur.cheng@oracle.com>
Co-authored-by: Bagatur <baskaryan@gmail.com>
2024-01-25 02:23:50 +00:00
|
|
|
from enum import Enum
|
|
|
|
from typing import Any, Dict, List, Mapping, Optional
|
|
|
|
|
|
|
|
from langchain_core.embeddings import Embeddings
|
|
|
|
from langchain_core.pydantic_v1 import BaseModel, Extra, root_validator
|
|
|
|
|
|
|
|
CUSTOM_ENDPOINT_PREFIX = "ocid1.generativeaiendpoint"
|
|
|
|
|
|
|
|
|
|
|
|
class OCIAuthType(Enum):
|
|
|
|
API_KEY = 1
|
|
|
|
SECURITY_TOKEN = 2
|
|
|
|
INSTANCE_PRINCIPAL = 3
|
|
|
|
RESOURCE_PRINCIPAL = 4
|
|
|
|
|
|
|
|
|
|
|
|
class OCIGenAIEmbeddings(BaseModel, Embeddings):
|
|
|
|
"""OCI embedding models.
|
|
|
|
|
|
|
|
To authenticate, the OCI client uses the methods described in
|
|
|
|
https://docs.oracle.com/en-us/iaas/Content/API/Concepts/sdk_authentication_methods.htm
|
|
|
|
|
|
|
|
The authentifcation method is passed through auth_type and should be one of:
|
|
|
|
API_KEY (default), SECURITY_TOKEN, INSTANCE_PRINCIPLE, RESOURCE_PRINCIPLE
|
|
|
|
|
|
|
|
Make sure you have the required policies (profile/roles) to
|
|
|
|
access the OCI Generative AI service. If a specific config profile is used,
|
|
|
|
you must pass the name of the profile (~/.oci/config) through auth_profile.
|
|
|
|
|
|
|
|
To use, you must provide the compartment id
|
|
|
|
along with the endpoint url, and model id
|
|
|
|
as named parameters to the constructor.
|
|
|
|
|
|
|
|
Example:
|
|
|
|
.. code-block:: python
|
|
|
|
|
|
|
|
from langchain.embeddings import OCIGenAIEmbeddings
|
|
|
|
|
|
|
|
embeddings = OCIGenAIEmbeddings(
|
|
|
|
model_id="MY_EMBEDDING_MODEL",
|
|
|
|
service_endpoint="https://inference.generativeai.us-chicago-1.oci.oraclecloud.com",
|
|
|
|
compartment_id="MY_OCID"
|
|
|
|
)
|
|
|
|
"""
|
|
|
|
|
|
|
|
client: Any #: :meta private:
|
|
|
|
|
|
|
|
service_models: Any #: :meta private:
|
|
|
|
|
|
|
|
auth_type: Optional[str] = "API_KEY"
|
|
|
|
"""Authentication type, could be
|
|
|
|
|
|
|
|
API_KEY,
|
|
|
|
SECURITY_TOKEN,
|
|
|
|
INSTANCE_PRINCIPLE,
|
|
|
|
RESOURCE_PRINCIPLE
|
|
|
|
|
|
|
|
If not specified, API_KEY will be used
|
|
|
|
"""
|
|
|
|
|
|
|
|
auth_profile: Optional[str] = "DEFAULT"
|
|
|
|
"""The name of the profile in ~/.oci/config
|
|
|
|
If not specified , DEFAULT will be used
|
|
|
|
"""
|
|
|
|
|
2024-02-05 19:22:06 +00:00
|
|
|
model_id: str = None # type: ignore[assignment]
|
community[minor]: Add OCI Generative AI integration (#16548)
<!-- Thank you for contributing to LangChain!
Please title your PR "<package>: <description>", where <package> is
whichever of langchain, community, core, experimental, etc. is being
modified.
Replace this entire comment with:
- **Description:** Adding Oracle Cloud Infrastructure Generative AI
integration. Oracle Cloud Infrastructure (OCI) Generative AI is a fully
managed service that provides a set of state-of-the-art, customizable
large language models (LLMs) that cover a wide range of use cases, and
which is available through a single API. Using the OCI Generative AI
service you can access ready-to-use pretrained models, or create and
host your own fine-tuned custom models based on your own data on
dedicated AI clusters.
https://docs.oracle.com/en-us/iaas/Content/generative-ai/home.htm
- **Issue:** None,
- **Dependencies:** OCI Python SDK,
- **Twitter handle:** we announce bigger features on Twitter. If your PR
gets announced, and you'd like a mention, we'll gladly shout you out!
Please make sure your PR is passing linting and testing before
submitting. Run `make format`, `make lint` and `make test` from the root
of the package you've modified to check this locally.
Passed
See contribution guidelines for more information on how to write/run
tests, lint, etc: https://python.langchain.com/docs/contributing/
If you're adding a new integration, please include:
1. a test for the integration, preferably unit tests that do not rely on
network access,
2. an example notebook showing its use. It lives in
`docs/docs/integrations` directory.
we provide unit tests. However, we cannot provide integration tests due
to Oracle policies that prohibit public sharing of api keys.
If no one reviews your PR within a few days, please @-mention one of
@baskaryan, @eyurtsev, @hwchase17.
-->
---------
Co-authored-by: Arthur Cheng <arthur.cheng@oracle.com>
Co-authored-by: Bagatur <baskaryan@gmail.com>
2024-01-25 02:23:50 +00:00
|
|
|
"""Id of the model to call, e.g., cohere.embed-english-light-v2.0"""
|
|
|
|
|
|
|
|
model_kwargs: Optional[Dict] = None
|
|
|
|
"""Keyword arguments to pass to the model"""
|
|
|
|
|
2024-02-05 19:22:06 +00:00
|
|
|
service_endpoint: str = None # type: ignore[assignment]
|
community[minor]: Add OCI Generative AI integration (#16548)
<!-- Thank you for contributing to LangChain!
Please title your PR "<package>: <description>", where <package> is
whichever of langchain, community, core, experimental, etc. is being
modified.
Replace this entire comment with:
- **Description:** Adding Oracle Cloud Infrastructure Generative AI
integration. Oracle Cloud Infrastructure (OCI) Generative AI is a fully
managed service that provides a set of state-of-the-art, customizable
large language models (LLMs) that cover a wide range of use cases, and
which is available through a single API. Using the OCI Generative AI
service you can access ready-to-use pretrained models, or create and
host your own fine-tuned custom models based on your own data on
dedicated AI clusters.
https://docs.oracle.com/en-us/iaas/Content/generative-ai/home.htm
- **Issue:** None,
- **Dependencies:** OCI Python SDK,
- **Twitter handle:** we announce bigger features on Twitter. If your PR
gets announced, and you'd like a mention, we'll gladly shout you out!
Please make sure your PR is passing linting and testing before
submitting. Run `make format`, `make lint` and `make test` from the root
of the package you've modified to check this locally.
Passed
See contribution guidelines for more information on how to write/run
tests, lint, etc: https://python.langchain.com/docs/contributing/
If you're adding a new integration, please include:
1. a test for the integration, preferably unit tests that do not rely on
network access,
2. an example notebook showing its use. It lives in
`docs/docs/integrations` directory.
we provide unit tests. However, we cannot provide integration tests due
to Oracle policies that prohibit public sharing of api keys.
If no one reviews your PR within a few days, please @-mention one of
@baskaryan, @eyurtsev, @hwchase17.
-->
---------
Co-authored-by: Arthur Cheng <arthur.cheng@oracle.com>
Co-authored-by: Bagatur <baskaryan@gmail.com>
2024-01-25 02:23:50 +00:00
|
|
|
"""service endpoint url"""
|
|
|
|
|
2024-02-05 19:22:06 +00:00
|
|
|
compartment_id: str = None # type: ignore[assignment]
|
community[minor]: Add OCI Generative AI integration (#16548)
<!-- Thank you for contributing to LangChain!
Please title your PR "<package>: <description>", where <package> is
whichever of langchain, community, core, experimental, etc. is being
modified.
Replace this entire comment with:
- **Description:** Adding Oracle Cloud Infrastructure Generative AI
integration. Oracle Cloud Infrastructure (OCI) Generative AI is a fully
managed service that provides a set of state-of-the-art, customizable
large language models (LLMs) that cover a wide range of use cases, and
which is available through a single API. Using the OCI Generative AI
service you can access ready-to-use pretrained models, or create and
host your own fine-tuned custom models based on your own data on
dedicated AI clusters.
https://docs.oracle.com/en-us/iaas/Content/generative-ai/home.htm
- **Issue:** None,
- **Dependencies:** OCI Python SDK,
- **Twitter handle:** we announce bigger features on Twitter. If your PR
gets announced, and you'd like a mention, we'll gladly shout you out!
Please make sure your PR is passing linting and testing before
submitting. Run `make format`, `make lint` and `make test` from the root
of the package you've modified to check this locally.
Passed
See contribution guidelines for more information on how to write/run
tests, lint, etc: https://python.langchain.com/docs/contributing/
If you're adding a new integration, please include:
1. a test for the integration, preferably unit tests that do not rely on
network access,
2. an example notebook showing its use. It lives in
`docs/docs/integrations` directory.
we provide unit tests. However, we cannot provide integration tests due
to Oracle policies that prohibit public sharing of api keys.
If no one reviews your PR within a few days, please @-mention one of
@baskaryan, @eyurtsev, @hwchase17.
-->
---------
Co-authored-by: Arthur Cheng <arthur.cheng@oracle.com>
Co-authored-by: Bagatur <baskaryan@gmail.com>
2024-01-25 02:23:50 +00:00
|
|
|
"""OCID of compartment"""
|
|
|
|
|
|
|
|
truncate: Optional[str] = "END"
|
|
|
|
"""Truncate embeddings that are too long from start or end ("NONE"|"START"|"END")"""
|
|
|
|
|
|
|
|
class Config:
|
|
|
|
"""Configuration for this pydantic object."""
|
|
|
|
|
|
|
|
extra = Extra.forbid
|
|
|
|
|
|
|
|
@root_validator()
|
|
|
|
def validate_environment(cls, values: Dict) -> Dict: # pylint: disable=no-self-argument
|
|
|
|
"""Validate that OCI config and python package exists in environment."""
|
|
|
|
|
|
|
|
# Skip creating new client if passed in constructor
|
|
|
|
if values["client"] is not None:
|
|
|
|
return values
|
|
|
|
|
|
|
|
try:
|
|
|
|
import oci
|
|
|
|
|
|
|
|
client_kwargs = {
|
|
|
|
"config": {},
|
|
|
|
"signer": None,
|
|
|
|
"service_endpoint": values["service_endpoint"],
|
|
|
|
"retry_strategy": oci.retry.DEFAULT_RETRY_STRATEGY,
|
|
|
|
"timeout": (10, 240), # default timeout config for OCI Gen AI service
|
|
|
|
}
|
|
|
|
|
|
|
|
if values["auth_type"] == OCIAuthType(1).name:
|
|
|
|
client_kwargs["config"] = oci.config.from_file(
|
|
|
|
profile_name=values["auth_profile"]
|
|
|
|
)
|
|
|
|
client_kwargs.pop("signer", None)
|
|
|
|
elif values["auth_type"] == OCIAuthType(2).name:
|
|
|
|
|
2024-02-05 19:22:06 +00:00
|
|
|
def make_security_token_signer(oci_config): # type: ignore[no-untyped-def]
|
community[minor]: Add OCI Generative AI integration (#16548)
<!-- Thank you for contributing to LangChain!
Please title your PR "<package>: <description>", where <package> is
whichever of langchain, community, core, experimental, etc. is being
modified.
Replace this entire comment with:
- **Description:** Adding Oracle Cloud Infrastructure Generative AI
integration. Oracle Cloud Infrastructure (OCI) Generative AI is a fully
managed service that provides a set of state-of-the-art, customizable
large language models (LLMs) that cover a wide range of use cases, and
which is available through a single API. Using the OCI Generative AI
service you can access ready-to-use pretrained models, or create and
host your own fine-tuned custom models based on your own data on
dedicated AI clusters.
https://docs.oracle.com/en-us/iaas/Content/generative-ai/home.htm
- **Issue:** None,
- **Dependencies:** OCI Python SDK,
- **Twitter handle:** we announce bigger features on Twitter. If your PR
gets announced, and you'd like a mention, we'll gladly shout you out!
Please make sure your PR is passing linting and testing before
submitting. Run `make format`, `make lint` and `make test` from the root
of the package you've modified to check this locally.
Passed
See contribution guidelines for more information on how to write/run
tests, lint, etc: https://python.langchain.com/docs/contributing/
If you're adding a new integration, please include:
1. a test for the integration, preferably unit tests that do not rely on
network access,
2. an example notebook showing its use. It lives in
`docs/docs/integrations` directory.
we provide unit tests. However, we cannot provide integration tests due
to Oracle policies that prohibit public sharing of api keys.
If no one reviews your PR within a few days, please @-mention one of
@baskaryan, @eyurtsev, @hwchase17.
-->
---------
Co-authored-by: Arthur Cheng <arthur.cheng@oracle.com>
Co-authored-by: Bagatur <baskaryan@gmail.com>
2024-01-25 02:23:50 +00:00
|
|
|
pk = oci.signer.load_private_key_from_file(
|
|
|
|
oci_config.get("key_file"), None
|
|
|
|
)
|
|
|
|
with open(
|
|
|
|
oci_config.get("security_token_file"), encoding="utf-8"
|
|
|
|
) as f:
|
|
|
|
st_string = f.read()
|
|
|
|
return oci.auth.signers.SecurityTokenSigner(st_string, pk)
|
|
|
|
|
|
|
|
client_kwargs["config"] = oci.config.from_file(
|
|
|
|
profile_name=values["auth_profile"]
|
|
|
|
)
|
|
|
|
client_kwargs["signer"] = make_security_token_signer(
|
|
|
|
oci_config=client_kwargs["config"]
|
|
|
|
)
|
|
|
|
elif values["auth_type"] == OCIAuthType(3).name:
|
|
|
|
client_kwargs[
|
|
|
|
"signer"
|
|
|
|
] = oci.auth.signers.InstancePrincipalsSecurityTokenSigner()
|
|
|
|
elif values["auth_type"] == OCIAuthType(4).name:
|
|
|
|
client_kwargs[
|
|
|
|
"signer"
|
|
|
|
] = oci.auth.signers.get_resource_principals_signer()
|
|
|
|
else:
|
|
|
|
raise ValueError("Please provide valid value to auth_type")
|
|
|
|
|
|
|
|
values["client"] = oci.generative_ai_inference.GenerativeAiInferenceClient(
|
|
|
|
**client_kwargs
|
|
|
|
)
|
|
|
|
|
|
|
|
except ImportError as ex:
|
|
|
|
raise ModuleNotFoundError(
|
|
|
|
"Could not import oci python package. "
|
|
|
|
"Please make sure you have the oci package installed."
|
|
|
|
) from ex
|
|
|
|
except Exception as e:
|
|
|
|
raise ValueError(
|
|
|
|
"Could not authenticate with OCI client. "
|
|
|
|
"Please check if ~/.oci/config exists. "
|
|
|
|
"If INSTANCE_PRINCIPLE or RESOURCE_PRINCIPLE is used, "
|
|
|
|
"Please check the specified "
|
|
|
|
"auth_profile and auth_type are valid."
|
|
|
|
) from e
|
|
|
|
|
|
|
|
return values
|
|
|
|
|
|
|
|
@property
|
|
|
|
def _identifying_params(self) -> Mapping[str, Any]:
|
|
|
|
"""Get the identifying parameters."""
|
|
|
|
_model_kwargs = self.model_kwargs or {}
|
|
|
|
return {
|
|
|
|
**{"model_kwargs": _model_kwargs},
|
|
|
|
}
|
|
|
|
|
|
|
|
def embed_documents(self, texts: List[str]) -> List[List[float]]:
|
|
|
|
"""Call out to OCIGenAI's embedding endpoint.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
texts: The list of texts to embed.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
List of embeddings, one for each text.
|
|
|
|
"""
|
|
|
|
from oci.generative_ai_inference import models
|
|
|
|
|
|
|
|
if self.model_id.startswith(CUSTOM_ENDPOINT_PREFIX):
|
|
|
|
serving_mode = models.DedicatedServingMode(endpoint_id=self.model_id)
|
|
|
|
else:
|
|
|
|
serving_mode = models.OnDemandServingMode(model_id=self.model_id)
|
|
|
|
|
|
|
|
invocation_obj = models.EmbedTextDetails(
|
|
|
|
serving_mode=serving_mode,
|
|
|
|
compartment_id=self.compartment_id,
|
|
|
|
truncate=self.truncate,
|
|
|
|
inputs=texts,
|
|
|
|
)
|
|
|
|
|
|
|
|
response = self.client.embed_text(invocation_obj)
|
|
|
|
|
|
|
|
return response.data.embeddings
|
|
|
|
|
|
|
|
def embed_query(self, text: str) -> List[float]:
|
|
|
|
"""Call out to OCIGenAI's embedding endpoint.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
text: The text to embed.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
Embeddings for the text.
|
|
|
|
"""
|
|
|
|
return self.embed_documents([text])[0]
|