DSpace Repository

Revieweval: an evaluation framework for ai-generated reviews

Show simple item record

dc.contributor.author Kumar, Dhruv
dc.date.accessioned 2025-04-25T06:41:33Z
dc.date.available 2025-04-25T06:41:33Z
dc.date.issued 2025
dc.identifier.uri https://arxiv.org/abs/2502.11736
dc.identifier.uri http://dspace.bits-pilani.ac.in:8080/jspui/handle/123456789/18785
dc.description.abstract The escalating volume of academic research, coupled with a shortage of qualified reviewers, necessitates innovative approaches to peer review. While large language model (LLMs) offer potential for automating this process, their current limitations include superficial critiques, hallucinations, and a lack of actionable insights. This research addresses these challenges by introducing a comprehensive evaluation framework for AI-generated reviews, that measures alignment with human evaluations, verifies factual accuracy, assesses analytical depth, and identifies actionable insights. We also propose a novel alignment mechanism that tailors LLM-generated reviews to the unique evaluation priorities of individual conferences and journals. To enhance the quality of these reviews, we introduce a self-refinement loop that iteratively optimizes the LLM's review prompts. Our framework establishes standardized metrics for evaluating AI-based review systems, thereby bolstering the reliability of AI-generated reviews in academic research. en_US
dc.language.iso en en_US
dc.subject Computer Science en_US
dc.subject Artificial Intelligence (AI) en_US
dc.subject Large language models (LLMs) en_US
dc.subject Alignment mechanism en_US
dc.title Revieweval: an evaluation framework for ai-generated reviews en_US
dc.type Preprint en_US


Files in this item

Files Size Format View

There are no files associated with this item.

This item appears in the following Collection(s)

Show simple item record

Search DSpace


Advanced Search

Browse

My Account