langchain/docs/use_cases
Vashisht Madhavan aa439ac2ff
Adding an in-context QA evaluation chain + chain of thought reasoning chain for improved accuracy (#2444)
Right now, eval chains require an answer for every question. It's
cumbersome to collect this ground truth so getting around this issue
with 2 things:

* Adding a context param in `ContextQAEvalChain` and simply evaluating
if the question is answered accurately from context
* Adding chain of though explanation prompting to improve the accuracy
of this w/o GT.

This also gets to feature parity with openai/evals which has the same
contextual eval w/o GT.

TODO in follow-up:
* Better prompt inheritance. No need for seperate prompt for CoT
reasoning. How can we merge them together

---------

Co-authored-by: Vashisht Madhavan <vashishtmadhavan@Vashs-MacBook-Pro.local>
2023-04-06 22:32:41 -07:00
..
evaluation Adding an in-context QA evaluation chain + chain of thought reasoning chain for improved accuracy (#2444) 2023-04-06 22:32:41 -07:00
apis.md Update apis.md (#2278) 2023-04-01 12:48:16 -07:00
chatbots.md Minor doc typo (#2492) 2023-04-06 12:41:40 -07:00
evaluation.rst big docs refactor (#1978) 2023-03-26 19:49:46 -07:00
extraction.md fix doc tags (#2019) 2023-03-26 21:43:51 -07:00
personal_assistants.md fix doc tags (#2019) 2023-03-26 21:43:51 -07:00
question_answering.md Harrison/site map (#2061) 2023-03-27 16:28:08 -07:00
summarization.md fix doc tags (#2019) 2023-03-26 21:43:51 -07:00
tabular.md fix doc tags (#2019) 2023-03-26 21:43:51 -07:00