Common Flaws in Running Human Evaluation Experiments in NLP (2024)
Attributed to:
ReproHum: Investigating Reproducibility of Human Evaluations in Natural Language Processing
funded by
EPSRC
Abstract
No abstract provided
Bibliographic Information
Digital Object Identifier: http://dx.doi.org/10.1162/coli_a_00508
Publication URI: http://dx.doi.org/10.1162/coli_a_00508
Type: Journal Article/Review
Parent Publication: Computational Linguistics
Issue: 2