forked from Archives/langchain
You cannot select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
aa439ac2ff
Right now, eval chains require an answer for every question. It's cumbersome to collect this ground truth so getting around this issue with 2 things: * Adding a context param in `ContextQAEvalChain` and simply evaluating if the question is answered accurately from context * Adding chain of though explanation prompting to improve the accuracy of this w/o GT. This also gets to feature parity with openai/evals which has the same contextual eval w/o GT. TODO in follow-up: * Better prompt inheritance. No need for seperate prompt for CoT reasoning. How can we merge them together --------- Co-authored-by: Vashisht Madhavan <vashishtmadhavan@Vashs-MacBook-Pro.local> |
1 year ago | |
---|---|---|
.. | ||
agents | 1 year ago | |
callbacks | 1 year ago | |
chains | 1 year ago | |
data | 2 years ago | |
docstore | 1 year ago | |
document_loader | 1 year ago | |
evaluation | 1 year ago | |
llms | 1 year ago | |
output_parsers | 1 year ago | |
prompts | 1 year ago | |
tools | 1 year ago | |
utilities | 1 year ago | |
__init__.py | 2 years ago | |
test_bash.py | 1 year ago | |
test_formatting.py | 2 years ago | |
test_python.py | 1 year ago | |
test_sql_database.py | 1 year ago | |
test_sql_database_schema.py | 1 year ago | |
test_text_splitter.py | 1 year ago |