Dr. Dave Howcroft D.Howcroft@napier.ac.uk
Associate
Most NLG is Low-Resource: here's what we can do about it
Howcroft, David M.; Gkatzia, Dimitra
Authors
Dr Dimitra Gkatzia D.Gkatzia@napier.ac.uk
Associate Professor
Abstract
Many domains and tasks in natural language generation (NLG) are inherently 'low-resource', where training data, tools and linguistic analyses are scarce. This poses a particular challenge to researchers and system developers in the era of machine-learning-driven NLG. In this position paper, we initially present the challenges researchers & developers often encounter when dealing with low-resource settings in NLG. We then argue that it is un-sustainable to collect large aligned datasets or build large language models from scratch for every possible domain due to cost, labour, and time constraints, so researching and developing methods and resources for low-resource settings is vital. We then discuss current approaches to low-resource NLG, followed by proposed solutions and promising avenues for future work in NLG for low-resource settings.
Citation
Howcroft, D. M., & Gkatzia, D. (2022, December). Most NLG is Low-Resource: here's what we can do about it. Presented at Workshop on Natural Language Generation, Evaluation, and Metrics (GEM), Abu Dhabi, UAE
Presentation Conference Type | Conference Paper (published) |
---|---|
Conference Name | Workshop on Natural Language Generation, Evaluation, and Metrics (GEM) |
Start Date | Dec 7, 2022 |
Acceptance Date | Oct 24, 2022 |
Publication Date | Dec 7, 2022 |
Deposit Date | Jun 9, 2023 |
Publicly Available Date | Jun 12, 2023 |
Publisher | Association for Computational Linguistics (ACL) |
Pages | 336-350 |
Book Title | Proceedings of the 2nd Workshop on Natural Language Generation, Evaluation, and Metrics (GEM) |
Publisher URL | https://aclanthology.org/2022.gem-1.29/ |
Files
Most NLG Is Low-Resource: Here's What We Can Do About It
(191 Kb)
PDF
Publisher Licence URL
http://creativecommons.org/licenses/by/4.0/
You might also like
How speakers adapt object descriptions to listeners under load
(2019)
Journal Article
OTTers: One-turn Topic Transitions for Open-Domain Dialogue
(-0001)
Presentation / Conference Contribution
Inducing Clause-Combining Rules: A Case Study with the SPaRKy Restaurant Corpus
(-0001)
Presentation / Conference Contribution
G-TUNA: a corpus of referring expressions in German, including duration information
(-0001)
Presentation / Conference Contribution
What happens if you treat ordinal ratings as interval data? Human evaluations in {NLP} are even more under-powered than you think
(-0001)
Presentation / Conference Contribution
Downloadable Citations
About Edinburgh Napier Research Repository
Administrator e-mail: repository@napier.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2024
Advanced Search