Please use this identifier to cite or link to this item: https://rfos.fon.bg.ac.rs/handle/123456789/2226
Full metadata record
DC FieldValueLanguage
dc.creatorPham, Ba
dc.creatorJovanović, Jelena
dc.creatorBagheri, Ebrahim
dc.creatorAntony, Jesmin
dc.creatorAshoor, Huda
dc.creatorNguyen, Tam T.
dc.creatorRios, Patricia
dc.creatorRobson, Reid
dc.creatorThomas, Sonia M.
dc.creatorWatt, Jennifer
dc.creatorStraus, Sharon E.
dc.creatorTricco, Andrea C.
dc.date.accessioned2023-05-12T11:36:55Z-
dc.date.available2023-05-12T11:36:55Z-
dc.date.issued2021
dc.identifier.issn2046-4053
dc.identifier.urihttps://rfos.fon.bg.ac.rs/handle/123456789/2226-
dc.description.abstractBackground: Current text mining tools supporting abstract screening in systematic reviews are not widely used, in part because they lack sensitivity and precision. We set out to develop an accessible, semi-automated "workflow" to conduct abstract screening for systematic reviews and other knowledge synthesis methods. Methods: We adopt widely recommended text-mining and machine-learning methods to (1) process title-abstracts into numerical training data; and (2) train a classification model to predict eligible abstracts. The predicted abstracts are screened by human reviewers for ("true") eligibility, and the newly eligible abstracts are used to identify similar abstracts, using near-neighbor methods, which are also screened. These abstracts, as well as their eligibility results, are used to update the classification model, and the above steps are iterated until no new eligible abstracts are identified. The workflow was implemented in R and evaluated using a systematic review of insulin formulations for type-1 diabetes (14,314 abstracts) and a scoping review of knowledge-synthesis methods (17,200 abstracts). Workflow performance was evaluated against the recommended practice of screening abstracts by 2 reviewers, independently. Standard measures were examined: sensitivity (inclusion of all truly eligible abstracts), specificity (exclusion of all truly ineligible abstracts), precision (inclusion of all truly eligible abstracts among all abstracts screened as eligible), F1-score (harmonic average of sensitivity and precision), and accuracy (correctly predicted eligible or ineligible abstracts). Workload reduction was measured as the hours the workflow saved, given only a subset of abstracts needed human screening. Results: With respect to the systematic and scoping reviews respectively, the workflow attained 88%/89% sensitivity, 99%/99% specificity, 71%/72% precision, an F1-score of 7996/79%, 98%/97% accuracy, 63%/55% workload reduction, with 12%/11% fewer abstracts for full-text retrieval and screening, and 0%/1.5% missed studies in the completed reviews. Conclusion: The workflow was a sensitive, precise, and efficient alternative to the recommended practice of screening abstracts with 2 reviewers. All eligible studies were identified in the first case, while 6 studies (1.5%) were missed in the second that would likely not impact the review's conclusions. We have described the workflow in language accessible to reviewers with limited exposure to natural language processing and machine learning, and have made the code available to reviewers.en
dc.publisherBMC, London
dc.relationTier 1 Canada Research Chair in Knowledge Translation
dc.relationSquires-Chalmers Chair for the Physician-in-Chief of Department of Medicine, St. Michael's Hospital
dc.relationUniversity of Toronto
dc.relationTier 2 Canada Research Chair in Knowledge Synthesis
dc.relationOntario Ministry of Research, Innovation, and Science
dc.relationNatural Sciences and Engineering Research Council (NSERC) Discovery Grants Program
dc.relationCanada Research Chairs Program
dc.relationNSERC Industrial Research Chairs Program
dc.rightsopenAccess
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/
dc.sourceSystematic Reviews
dc.subjectText miningen
dc.subjectSystematic reviewen
dc.subjectScoping reviewen
dc.subjectNatural language processingen
dc.subjectMachine learningen
dc.subjectClassification modelen
dc.subjectAutomationen
dc.subjectAbstract screeningen
dc.titleText mining to support abstract screening for knowledge syntheses: a semi-automated workflowen
dc.typearticle
dc.rights.licenseBY
dc.citation.issue1
dc.citation.other10(1): -
dc.citation.rankM22
dc.citation.volume10
dc.identifier.doi10.1186/s13643-021-01700-x
dc.identifier.fulltexthttp://prototype2.rcub.bg.ac.rs/bitstream/id/784/2222.pdf
dc.identifier.pmid34039433
dc.identifier.rcubconv_2501
dc.identifier.scopus2-s2.0-85106886290
dc.identifier.wos000657785900002
dc.type.versionpublishedVersion
item.cerifentitytypePublications-
item.fulltextWith Fulltext-
item.grantfulltextopen-
item.openairetypearticle-
item.openairecristypehttp://purl.org/coar/resource_type/c_18cf-
Appears in Collections:Radovi istraživača / Researchers’ publications
Files in This Item:
File Description SizeFormat 
2222.pdf1.38 MBAdobe PDFThumbnail
View/Open
Show simple item record

SCOPUSTM   
Citations

32
checked on Nov 17, 2025

Google ScholarTM

Check

Altmetric


This item is licensed under a Creative Commons License Creative Commons