Login (DCU Staff Only)
Login (DCU Staff Only)

DORAS | DCU Research Repository

Explore open access research and scholarly works from DCU

Advanced Search

The human evaluation datasheet: a template for recording details of human evaluation experiments in NLP

Shimorina, Anastasia and Belz, Anya orcid logoORCID: 0000-0002-0552-8096 (2022) The human evaluation datasheet: a template for recording details of human evaluation experiments in NLP. In: 2nd Workshop on Human Evaluation of NLP Systems, 27 May 2022, Dublin, Ireland. ISBN 9781713867395

Abstract
This paper presents the Human Evaluation Datasheet (HEDS), a template for recording the details of individual human evaluation experiments in Natural Language Processing (NLP), and reports on first experience of researchers using HEDS sheets in practice. Originally taking inspiration from seminal papers by Bender and Friedman (2018), Mitchell et al. (2019), and Gebru et al. (2020), HEDS facilitates the recording of properties of human evaluations in sufficient detail, and with sufficient standardisation, to support comparability, meta-evaluation, and reproducibility assessments for human evaluations. These are crucial for scientifically principled evaluation, but the overhead of completing a detailed datasheet is substantial, and we discuss possible ways of addressing this and other issues observed in practice.
Metadata
Item Type:Conference or Workshop Item (Paper)
Event Type:Workshop
Refereed:Yes
Subjects:Computer Science > Computational linguistics
DCU Faculties and Centres:DCU Faculties and Schools > Faculty of Engineering and Computing > School of Computing
Research Institutes and Centres > ADAPT
Published in: Belz, Anya, Popović, Maja, Reiter, Ehud and Shimorina, Anastasia, (eds.) Proceedings of the 2nd Workshop on Human Evaluation of NLP Systems (HumEval). . Association for Computing Machinery (ACM). ISBN 9781713867395
Publisher:Association for Computing Machinery (ACM)
Official URL:https://doi.org/10.18653/v1/2022.humeval-1.6
Copyright Information:© 2022 ACL
ID Code:28649
Deposited On:06 Jul 2023 09:49 by Anya Belz . Last Modified 06 Jul 2023 09:49
Documents

Full text available as:

[thumbnail of 2022.humeval-1.6.pdf]
Preview
PDF - Requires a PDF viewer such as GSview, Xpdf or Adobe Acrobat Reader
Creative Commons: Attribution-Noncommercial 4.0
281kB
Metrics

Altmetric Badge

Dimensions Badge

Downloads

Downloads

Downloads per month over past year

Archive Staff Only: edit this record