DSpace logo

Please use this identifier to cite or link to this item: http://dspace.bits-pilani.ac.in:8080/jspui/handle/123456789/18785
Full metadata record
DC FieldValueLanguage
dc.contributor.authorKumar, Dhruv-
dc.date.accessioned2025-04-25T06:41:33Z-
dc.date.available2025-04-25T06:41:33Z-
dc.date.issued2025-
dc.identifier.urihttps://arxiv.org/abs/2502.11736-
dc.identifier.urihttp://dspace.bits-pilani.ac.in:8080/jspui/handle/123456789/18785-
dc.description.abstractThe escalating volume of academic research, coupled with a shortage of qualified reviewers, necessitates innovative approaches to peer review. While large language model (LLMs) offer potential for automating this process, their current limitations include superficial critiques, hallucinations, and a lack of actionable insights. This research addresses these challenges by introducing a comprehensive evaluation framework for AI-generated reviews, that measures alignment with human evaluations, verifies factual accuracy, assesses analytical depth, and identifies actionable insights. We also propose a novel alignment mechanism that tailors LLM-generated reviews to the unique evaluation priorities of individual conferences and journals. To enhance the quality of these reviews, we introduce a self-refinement loop that iteratively optimizes the LLM's review prompts. Our framework establishes standardized metrics for evaluating AI-based review systems, thereby bolstering the reliability of AI-generated reviews in academic research.en_US
dc.language.isoenen_US
dc.subjectComputer Scienceen_US
dc.subjectArtificial Intelligence (AI)en_US
dc.subjectLarge language models (LLMs)en_US
dc.subjectAlignment mechanismen_US
dc.titleRevieweval: an evaluation framework for ai-generated reviewsen_US
dc.typePreprinten_US
Appears in Collections:Department of Computer Science and Information Systems

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.