NLPExplorer
Papers
Venues
Authors
Authors Timeline
Field of Study
URLs
ACL N-gram Stats
TweeNLP
API
Team
HumEval - 2021
Total Papers:- 16
Total Papers accross all years:- 71
Total Citations :- 0
1
2
»
The Great Misalignment Problem in Human Evaluation of NLP Methods
Mika Hämäläinen |
Khalid Alnajjar |
Trading Off Diversity and Quality in Natural Language Generation
Hugh Zhang |
Daniel Duckworth |
Daphne Ippolito |
Arvind Neelakantan |
Interrater Disagreement Resolution: A Systematic Procedure to Reach Consensus in Annotation Tasks
Yvette Oortwijn |
Thijs Ossenkoppele |
Arianna Betti |
It’s Commonsense, isn’t it? Demystifying Human Evaluations in Commonsense-Enhanced NLG Systems
Miruna-Adriana Clinciu |
Dimitra Gkatzia |
Saad Mahamood |
Is This Translation Error Critical?: Classification-Based Human and Automatic Machine Translation Evaluation Focusing on Critical Errors
Katsuhito Sudoh |
Kosuke Takahashi |
Satoshi Nakamura |
Towards Document-Level Human MT Evaluation: On the Issues of Annotator Agreement, Effort and Misevaluation
Sheila Castilho |
Towards Objectively Evaluating the Quality of Generated Medical Summaries
Francesco Moramarco |
Damir Juric |
Aleksandar Savkov |
Ehud Reiter |
Eliciting Explicit Knowledge From Domain Experts in Direct Intrinsic Evaluation of Word Embeddings for Specialized Domains
Goya van Boven |
Jelke Bloem |
Proceedings of the Workshop on Human Evaluation of NLP Systems (HumEval)
Anya Belz |
Shubham Agarwal |
Yvette Graham |
Ehud Reiter |
Anastasia Shimorina |
Estimating Subjective Crowd-Evaluations as an Additional Objective to Improve Natural Language Generation
Jakob Nyberg |
Maike Paetzel |
Ramesh Manuvinakurike |
A Case Study of Efficacy and Challenges in Practical Human-in-Loop Evaluation of NLP Systems Using Checklist
Shaily Bhatt |
Rahul Jain |
Sandipan Dandapat |
Sunayana Sitaram |
A View From the Crowd: Evaluation Challenges for Time-Offset Interaction Applications
Alberto Chierici |
Nizar Habash |
A Preliminary Study on Evaluating Consultation Notes With Post-Editing
Francesco Moramarco |
Alex Papadopoulos Korfiatis |
Aleksandar Savkov |
Ehud Reiter |
Detecting Post-Edited References and Their Effect on Human Evaluation
Věra Kloudová |
Ondřej Bojar |
Martin Popel |
On User Interfaces for Large-Scale Document-Level Human Evaluation of Machine Translation Outputs
Roman Grundkiewicz |
Marcin Junczys-Dowmunt |
Christian Federmann |
Tom Kocmi |
Conference Topic Distribution
Linguistic
Task
Approach
Language
Dataset
Conference Citation Distribution
Conference Papers have no Citations yet
Topics