diff --git a/libs/partners/openai/langchain_openai/embeddings/azure.py b/libs/partners/openai/langchain_openai/embeddings/azure.py index f162c18ecd..de53c8058c 100644 --- a/libs/partners/openai/langchain_openai/embeddings/azure.py +++ b/libs/partners/openai/langchain_openai/embeddings/azure.py @@ -102,14 +102,6 @@ class AzureOpenAIEmbeddings(OpenAIEmbeddings): values["azure_ad_token"] = ( convert_to_secret_str(azure_ad_token) if azure_ad_token else None ) - # Azure OpenAI embedding models allow a maximum of 2048 texts - # at a time in each batch - # See: https://learn.microsoft.com/en-us/azure/ai-services/openai/how-to/embeddings?tabs=console#best-practices - if values["chunk_size"] > 2048: - raise ValueError( - "Azure OpenAI embeddings only allow a maximum of 2048 texts at a time " - "in each batch." - ) # For backwards compatibility. Before openai v1, no distinction was made # between azure_endpoint and base_url (openai_api_base). openai_api_base = values["openai_api_base"]