From 57e028549abfd4649076a49e61d6ab9cb3478839 Mon Sep 17 00:00:00 2001 From: Rafal Wojdyla Date: Sat, 29 Apr 2023 05:18:05 +0100 Subject: [PATCH] Expose kwargs in `LLMChainExtractor.from_llm` (#3748) Re: https://github.com/hwchase17/langchain/issues/3747 --- langchain/retrievers/document_compressors/chain_extract.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/langchain/retrievers/document_compressors/chain_extract.py b/langchain/retrievers/document_compressors/chain_extract.py index 6f638559..175dd76d 100644 --- a/langchain/retrievers/document_compressors/chain_extract.py +++ b/langchain/retrievers/document_compressors/chain_extract.py @@ -69,9 +69,10 @@ class LLMChainExtractor(BaseDocumentCompressor): llm: BaseLanguageModel, prompt: Optional[PromptTemplate] = None, get_input: Optional[Callable[[str, Document], str]] = None, + llm_chain_kwargs: Optional[dict] = None, ) -> "LLMChainExtractor": """Initialize from LLM.""" _prompt = prompt if prompt is not None else _get_default_chain_prompt() _get_input = get_input if get_input is not None else default_get_input - llm_chain = LLMChain(llm=llm, prompt=_prompt) + llm_chain = LLMChain(llm=llm, prompt=_prompt, **(llm_chain_kwargs or {})) return cls(llm_chain=llm_chain, get_input=_get_input)