We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
TO BE DEFINED
@Lanture1064 @bjwswang
Our current RAG solution flow :
Based on our research ,we decide to use this evaluation framework https://github.com/explodinggradients/ragas
Subtasks:
@0xff-dev @bjwswang
For definitions:
For task runner:
For apiserver:
The text was updated successfully, but these errors were encountered:
@Lanture1064 pls do further investigation.
Sorry, something went wrong.
here is another project related to evaluate: https://github.com/promptfoo/promptfoo
Some other thoughts:
Some other thoughts: 支持 Prompt 在不同 LLM 下的评估,生成测试报告 RAG 评估、RAG Question Generation 自动生成问题,分析问题质量,过滤掉相似度不高的问题 评估指标:检索评估 - Hit Rate、MRR,回答评估 - 公正性、相关性、一致性等
支持 Prompt 在不同 LLM 下的评估,生成测试报告
This story can be closed. For other features like support tektonci and other llms ,we can track in individual issues.
0xff-dev
bjwswang
Lanture1064
No branches or pull requests
Overall workflow
TO BE DEFINED
Evaluation Types
RAG Evaluation
@Lanture1064 @bjwswang
Our current RAG solution flow :
Based on our research ,we decide to use this evaluation framework https://github.com/explodinggradients/ragas
Subtasks:
Evaluation Lifecycle management
@0xff-dev @bjwswang
For definitions:
For task runner:
For apiserver:
Overall Workflow
The text was updated successfully, but these errors were encountered: