Skip to content

SadaDC91/RAG_Evaluation_Task

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 

Repository files navigation

RAG_Evaluation_Task

A short program that performs a basic RAG evaluation

Retrieval-Augmented Generation (RAG) Evaluation

This model generation model. Accurate retrieval is critical for generating contextually relevant responses.

Types of RAG Evaluation

  1. Retrieval Evaluation:
    • Focuses on evaluating the relevance of retrieved documents.
    • Metrics: Precision@k, Recall@k, Mean Reciprocal Rank (MRR), etc.
  2. Generation Evaluation:
    • Measures the quality of generated text based on retrieved documents.
    • Metrics: BLEU, ROUGE, BERTScore, etc.
  3. End-to-End Evaluation:
    • Directly evaluates the RAG system's output for correctness and relevance.
    • Metrics: Human evaluation or task-specific benchmarks.

Chosen Evaluation for my model here: Retrieval Evaluation

The evaluation here focuses on retrieval because the quality of retrieved documents is foundational to the RAG framework's success. Precision@1 and MRR are used as metrics for this evaluation.

About

A short program that performs a basic RAG evaluation

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages