Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

RAG Evaluation Scripts #11

Open
qiaoruiyt opened this issue Jan 10, 2025 · 1 comment
Open

RAG Evaluation Scripts #11

qiaoruiyt opened this issue Jan 10, 2025 · 1 comment

Comments

@qiaoruiyt
Copy link

Hi authors!

Thank you for curating the nice dataset! May I ask whether it is possible to release the RAG evaluation scripts used to generate "Table 4: Question-answering results with different retrievers." in v3 of the paper on Arxiv? We are trying to better understand how IR improves RAG subsequently. Thank you!

@Frankgu3528
Copy link

I tried to implement the RAG generation and evaluation process on my own and found the “None” results were almost always better then results with retrieved document. I tried Llama-3.1-70B-Instruct and Gemini-2.0 as generation and evaluation model, tried different retrievers but the results were the same. I would greatly appreciate it if you could share the RAG evaluation scripts. Thanks in advance!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants