|
|
@ -73,7 +73,12 @@ class BaseLLM(BaseModel, ABC):
|
|
|
|
self.callback_manager.on_llm_start(
|
|
|
|
self.callback_manager.on_llm_start(
|
|
|
|
{"name": self.__class__.__name__}, prompts
|
|
|
|
{"name": self.__class__.__name__}, prompts
|
|
|
|
)
|
|
|
|
)
|
|
|
|
output = self._generate(prompts, stop=stop)
|
|
|
|
try:
|
|
|
|
|
|
|
|
output = self._generate(prompts, stop=stop)
|
|
|
|
|
|
|
|
except Exception as e:
|
|
|
|
|
|
|
|
if self.verbose:
|
|
|
|
|
|
|
|
self.callback_manager.on_llm_error(e)
|
|
|
|
|
|
|
|
raise e
|
|
|
|
if self.verbose:
|
|
|
|
if self.verbose:
|
|
|
|
self.callback_manager.on_llm_end(output)
|
|
|
|
self.callback_manager.on_llm_end(output)
|
|
|
|
return output
|
|
|
|
return output
|
|
|
@ -90,11 +95,18 @@ class BaseLLM(BaseModel, ABC):
|
|
|
|
else:
|
|
|
|
else:
|
|
|
|
missing_prompts.append(prompt)
|
|
|
|
missing_prompts.append(prompt)
|
|
|
|
missing_prompt_idxs.append(i)
|
|
|
|
missing_prompt_idxs.append(i)
|
|
|
|
self.callback_manager.on_llm_start(
|
|
|
|
if self.verbose:
|
|
|
|
{"name": self.__class__.__name__}, missing_prompts
|
|
|
|
self.callback_manager.on_llm_start(
|
|
|
|
)
|
|
|
|
{"name": self.__class__.__name__}, missing_prompts
|
|
|
|
new_results = self._generate(missing_prompts, stop=stop)
|
|
|
|
)
|
|
|
|
self.callback_manager.on_llm_end(new_results)
|
|
|
|
try:
|
|
|
|
|
|
|
|
new_results = self._generate(missing_prompts, stop=stop)
|
|
|
|
|
|
|
|
except Exception as e:
|
|
|
|
|
|
|
|
if self.verbose:
|
|
|
|
|
|
|
|
self.callback_manager.on_llm_error(e)
|
|
|
|
|
|
|
|
raise e
|
|
|
|
|
|
|
|
if self.verbose:
|
|
|
|
|
|
|
|
self.callback_manager.on_llm_end(new_results)
|
|
|
|
for i, result in enumerate(new_results.generations):
|
|
|
|
for i, result in enumerate(new_results.generations):
|
|
|
|
existing_prompts[missing_prompt_idxs[i]] = result
|
|
|
|
existing_prompts[missing_prompt_idxs[i]] = result
|
|
|
|
prompt = prompts[i]
|
|
|
|
prompt = prompts[i]
|
|
|
|