|
|
|
@ -435,7 +435,6 @@ class BaseLLM(BaseLanguageModel[str], ABC):
|
|
|
|
|
options=options,
|
|
|
|
|
name=config.get("run_name"),
|
|
|
|
|
run_id=config.pop("run_id", None),
|
|
|
|
|
batch_size=1,
|
|
|
|
|
)
|
|
|
|
|
generation: Optional[GenerationChunk] = None
|
|
|
|
|
try:
|
|
|
|
@ -496,7 +495,6 @@ class BaseLLM(BaseLanguageModel[str], ABC):
|
|
|
|
|
options=options,
|
|
|
|
|
name=config.get("run_name"),
|
|
|
|
|
run_id=config.pop("run_id", None),
|
|
|
|
|
batch_size=1,
|
|
|
|
|
)
|
|
|
|
|
generation: Optional[GenerationChunk] = None
|
|
|
|
|
try:
|
|
|
|
@ -793,7 +791,6 @@ class BaseLLM(BaseLanguageModel[str], ABC):
|
|
|
|
|
invocation_params=params,
|
|
|
|
|
options=options,
|
|
|
|
|
name=run_name,
|
|
|
|
|
batch_size=len(prompts),
|
|
|
|
|
run_id=run_id_,
|
|
|
|
|
)[0]
|
|
|
|
|
for callback_manager, prompt, run_name, run_id_ in zip(
|
|
|
|
@ -812,7 +809,6 @@ class BaseLLM(BaseLanguageModel[str], ABC):
|
|
|
|
|
invocation_params=params,
|
|
|
|
|
options=options,
|
|
|
|
|
name=run_name_list[idx],
|
|
|
|
|
batch_size=len(missing_prompts),
|
|
|
|
|
)[0]
|
|
|
|
|
for idx in missing_prompt_idxs
|
|
|
|
|
]
|
|
|
|
@ -1006,7 +1002,6 @@ class BaseLLM(BaseLanguageModel[str], ABC):
|
|
|
|
|
invocation_params=params,
|
|
|
|
|
options=options,
|
|
|
|
|
name=run_name,
|
|
|
|
|
batch_size=len(prompts),
|
|
|
|
|
run_id=run_id_,
|
|
|
|
|
)
|
|
|
|
|
for callback_manager, prompt, run_name, run_id_ in zip(
|
|
|
|
@ -1032,7 +1027,6 @@ class BaseLLM(BaseLanguageModel[str], ABC):
|
|
|
|
|
invocation_params=params,
|
|
|
|
|
options=options,
|
|
|
|
|
name=run_name_list[idx],
|
|
|
|
|
batch_size=len(missing_prompts),
|
|
|
|
|
)
|
|
|
|
|
for idx in missing_prompt_idxs
|
|
|
|
|
]
|
|
|
|
|