Abstract
Evaluation has long been a central concern in
NLP, and transparent reporting practices are
more critical than ever in today’s landscape of
rapidly released open-access models. Draw-
ing on a survey of recent work on evaluation
and documentation, we identify three persis-
tent shortcomings in current reporting prac-
tices: reproducibility, accessibility, and gover-
nance. We argue that existing standardization
efforts remain insufficient and introduce Evalu-
ation Disclosure Cards (EvalCards) as a path
forward. EvalCards are designed to enhance
transparency for both researchers and practi-
tioners while providing a practical foundation
to meet emerging governance requirements.
NLP, and transparent reporting practices are
more critical than ever in today’s landscape of
rapidly released open-access models. Draw-
ing on a survey of recent work on evaluation
and documentation, we identify three persis-
tent shortcomings in current reporting prac-
tices: reproducibility, accessibility, and gover-
nance. We argue that existing standardization
efforts remain insufficient and introduce Evalu-
ation Disclosure Cards (EvalCards) as a path
forward. EvalCards are designed to enhance
transparency for both researchers and practi-
tioners while providing a practical foundation
to meet emerging governance requirements.
| Original language | English |
|---|---|
| DOIs | |
| Publication status | Published - 2025 |
Fields of science
- 102003 Image processing
- 202002 Audiovisual media
- 102001 Artificial intelligence
- 102015 Information systems
- 102 Computer Sciences
JKU Focus areas
- Digital Transformation
Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver