- Annotation Files
- Contains human annotation for micro and macro evaluation.
- LLM Responses
- This folder contains responses generated by three LLMs (i.e., PalM2, GPT-3.5, and Llama2) for four prompt levels. Here, LLM1, LLM2, and LLM3 represents
PalM2
,GPT-3.5
, andLlama2
. The responses are in the columnsres_prompt1
,res_prompt2
,res_prompt3
, andres_prompt4
.
- This folder contains responses generated by three LLMs (i.e., PalM2, GPT-3.5, and Llama2) for four prompt levels. Here, LLM1, LLM2, and LLM3 represents
- GPT Ratings
- Contains the ratings given by
GPT-4o
on the LLM generated responses. The human and GPT ratings are merged together in the files.GPT_rating_LLMx
andMacro_GPT_score
contains micro and macro evaluation ratings of both human andGPT-4o
respectively. corr_average_likert_ceil
file contains the correlations between human and GPT-4o in micro-evaluation for each LLMs.
- Contains the ratings given by
- Meta-Review Data
- The curated dataset used for the meta-review generation.
Review1
,Review2
, andReview3
columns have the peer reviews andMeta_Review
column contain the meta-reviews written by humans.
- The curated dataset used for the meta-review generation.
-
Notifications
You must be signed in to change notification settings - Fork 0
BridgeAI-Lab/LLM-as-Meta-Reviewer
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
About
[NAACL'25] Dataset and Evaluation Code for LLMs as Meta-Reviewers’ Assistants: A Case Study