Babatunde Kazeem Olorisade
Reproducibility of studies on text mining for citation screening in systematic reviews: evaluation and checklist
Olorisade, Babatunde Kazeem; Brereton, Pearl; Andras, Peter
Authors
Pearl Brereton
Prof Peter Andras P.Andras@napier.ac.uk
Dean of School of Computing Engineering and the Built Environment
Abstract
Context
Independent validation of published scientific results through study replication is a pre-condition for accepting the validity of such results. In computation research, full replication is often unrealistic for independent results validation, therefore, study reproduction has been justified as the minimum acceptable standard to evaluate the validity of scientific claims. The application of text mining techniques to citation screening in the context of systematic literature reviews is a relatively young and growing computational field with high relevance for software engineering, medical research and other fields. However, there is little work so far on reproduction studies in the field.
Objective
In this paper, we investigate the reproducibility of studies in this area based on information contained in published articles and we propose reporting guidelines that could improve reproducibility.
Methods
The study was approached in two ways. Initially we attempted to reproduce results from six studies, which were based on the same raw dataset. Then, based on this experience, we identified steps considered essential to successful reproduction of text mining experiments and characterized them to measure how reproducible is a study given the information provided on these steps. 33 articles were systematically assessed for reproducibility using this approach.
Results
Our work revealed that it is currently difficult if not impossible to independently reproduce the results published in any of the studies investigated. The lack of information about the datasets used limits reproducibility of about 80% of the studies assessed. Also, information about the machine learning algorithms is inadequate in about 27% of the papers. On the plus side, the third party software tools used are mostly free and available.
Conclusions
The reproducibility potential of most of the studies can be significantly improved if more attention is paid to information provided on the datasets used, how they were partitioned and utilized, and how any randomization was controlled. We introduce a checklist of information that needs to be provided in order to ensure that a published study can be reproduced.
Citation
Olorisade, B. K., Brereton, P., & Andras, P. (2017). Reproducibility of studies on text mining for citation screening in systematic reviews: evaluation and checklist. Journal of Biomedical Informatics, 73, 1-13. https://doi.org/10.1016/j.jbi.2017.07.010
Journal Article Type | Article |
---|---|
Acceptance Date | Jul 10, 2017 |
Online Publication Date | Jul 12, 2017 |
Publication Date | 2017-09 |
Deposit Date | Nov 8, 2021 |
Journal | Journal of biomedical informatics |
Print ISSN | 1532-0464 |
Publisher | Elsevier |
Peer Reviewed | Peer Reviewed |
Volume | 73 |
Pages | 1-13 |
DOI | https://doi.org/10.1016/j.jbi.2017.07.010 |
Keywords | Citation screening, Systematic review, Reproducibility, Text mining, Reproducible research |
Public URL | http://researchrepository.napier.ac.uk/Output/2808928 |
You might also like
A review of privacy-preserving federated learning for the Internet-of-Things
(2021)
Book Chapter
Amnesia: Neuropsychological Interpretation and Artificial Neural Network Simulation
(1998)
Journal Article
Neural activity pattern systems
(2004)
Journal Article
Medical research funding may have over-expanded and be due for collapse
(2005)
Journal Article
Downloadable Citations
About Edinburgh Napier Research Repository
Administrator e-mail: repository@napier.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2024
Advanced Search