langchain/docs/use_cases
Graham Neubig 31303d0b11
Added other evaluation metrics for data-augmented QA (#1521)
This PR adds additional evaluation metrics for data-augmented QA,
resulting in a report like this at the end of the notebook:

![Screen Shot 2023-03-08 at 8 53 23
AM](https://user-images.githubusercontent.com/398875/223731199-8eb8e77f-5ff3-40a2-a23e-f3bede623344.png)

The score calculation is based on the
[Critique](https://docs.inspiredco.ai/critique/) toolkit, an API-based
toolkit (like OpenAI) that has minimal dependencies, so it should be
easy for people to run if they choose.

The code could further be simplified by actually adding a chain that
calls Critique directly, but that probably should be saved for another
PR if necessary. Any comments or change requests are welcome!
2023-03-08 20:41:03 -08:00
..
evaluation Added other evaluation metrics for data-augmented QA (#1521) 2023-03-08 20:41:03 -08:00
agents.md docs: fix spelling typo (#934) 2023-02-08 11:13:35 -08:00
chatbots.md bump to 0058 (#556) 2023-01-06 07:58:38 -08:00
combine_docs.md Harrison/updating docs (#1196) 2023-02-20 22:54:26 -08:00
evaluation.rst Feature: linkcheck-action (#534) (#542) 2023-01-04 21:39:50 -08:00
generate_examples.ipynb Docs refactor (#480) 2023-01-02 08:24:09 -08:00
model_laboratory.ipynb Docs refactor (#480) 2023-01-02 08:24:09 -08:00
question_answering.md bump version to 0.0.95 (#1324) 2023-02-27 07:45:54 -08:00
summarization.md Fix link in summarization.md (#1344) 2023-02-28 18:58:12 -08:00