Lewis N. Watson
Reproducing Human Evaluation of Meaning Preservation in Paraphrase Generation
Watson, Lewis N.; Gkatzia, Dimitra
Abstract
Reproducibility is a cornerstone of scientific research, ensuring the reliability and generalisability of findings. The ReproNLP Shared Task on Reproducibility of Evaluations in NLP aims to assess the reproducibility of human evaluation studies. This paper presents a reproduction study of the human evaluation experiment presented in "Hierarchical Sketch Induction for Paraphrase Generation" by Hosking et al. (2022). The original study employed a human evaluation on Amazon Mechanical Turk, assessing the quality of paraphrases generated by their proposed model using three criteria: meaning preservation, fluency, and dissimilarity. In our reproduction study, we focus on the meaning preservation criterion and utilise the Prolific platform for participant recruitment, following the ReproNLP challenge’s common approach to reproduction. We discuss the methodology, results, and implications of our reproduction study, comparing them to the original findings. Our findings contribute to the understanding of reproducibility in NLP research and highlights the potential impact of platform changes and evaluation criteria on the reproducibility of human evaluation studies.
Citation
Watson, L. N., & Gkatzia, D. (in press). Reproducing Human Evaluation of Meaning Preservation in Paraphrase Generation.
Conference Name | HumEval2024 at LREC-COLING 2024 |
---|---|
Conference Location | Turin, Italy |
Start Date | May 21, 2024 |
End Date | May 21, 2024 |
Acceptance Date | Apr 9, 2024 |
Deposit Date | Apr 11, 2024 |
Public URL | http://researchrepository.napier.ac.uk/Output/3590573 |
Related Public URLs | https://humeval.github.io/ |
You might also like
TaskMaster: A Novel Cross-platform Task-based Spoken Dialogue System for Human-Robot Interaction
(2023)
Conference Proceeding
Building a dual dataset of text-and image-grounded conversations and summarisation in Gàidhlig (Scottish Gaelic)
(2023)
Conference Proceeding
enunlg: a Python library for reproducible neural data-to-text experimentation
(2023)
Conference Proceeding
Downloadable Citations
About Edinburgh Napier Research Repository
Administrator e-mail: repository@napier.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2024
Advanced Search