Does ChatGPT Ignore Article Retractions and Other Reliability Concerns?

dc.contributor.authorThelwall, Mike
dc.contributor.authorLehtisaari, Marianna
dc.contributor.authorKatsirea, Irini
dc.contributor.authorHolmberg, Kim
dc.contributor.authorZheng, Er‐Te
dc.contributor.organizationfi=taloussosiologia|en=Economic Sociology|
dc.contributor.organization-code1.2.246.10.2458963.20.82939713796
dc.converis.publication-id499720452
dc.converis.urlhttps://research.utu.fi/converis/portal/Publication/499720452
dc.date.accessioned2026-01-21T12:09:17Z
dc.date.available2026-01-21T12:09:17Z
dc.description.abstractLarge language models (LLMs) like ChatGPT seem to be increasingly used for information seeking and analysis, including to support academic literature reviews. To test whether the results might sometimes include retracted research, we identified 217 retracted or otherwise concerning academic studies with high altmetric scores and asked ChatGPT 4o-mini to evaluate their quality 30 times each. Surprisingly, none of its 6510 reports mentioned that the articles were retracted or had relevant errors, and it gave 190 relatively high scores (world leading, internationally excellent, or close). The 27 articles with the lowest scores were mostly accused of being weak, although the topic (but not the article) was described as controversial in five cases (e.g., about hydroxychloroquine for COVID-19). In a follow-up investigation, 61 claims were extracted from retracted articles from the set, and ChatGPT 4o-mini was asked 10 times whether each was true. It gave a definitive yes or a positive response two-thirds of the time, including for at least one statement that had been shown to be false over a decade ago. The results therefore emphasise, from an academic knowledge perspective, the importance of verifying information from LLMs when using them for information seeking or analysis.
dc.identifier.eissn1741-4857
dc.identifier.jour-issn0953-1513
dc.identifier.olddbid212171
dc.identifier.oldhandle10024/195189
dc.identifier.urihttps://www.utupub.fi/handle/11111/40356
dc.identifier.urlhttps://doi.org/10.1002/leap.2018
dc.identifier.urnURN:NBN:fi-fe202601216605
dc.language.isoen
dc.okm.affiliatedauthorLehtisaari, Marianna
dc.okm.affiliatedauthorHolmberg, Kim
dc.okm.discipline113 Computer and information sciencesen_GB
dc.okm.discipline5141 Sociologyen_GB
dc.okm.discipline113 Tietojenkäsittely ja informaatiotieteetfi_FI
dc.okm.discipline5141 Sosiologiafi_FI
dc.okm.internationalcopublicationinternational co-publication
dc.okm.internationalityInternational publication
dc.okm.typeA1 ScientificArticle
dc.publisherWiley
dc.publisher.countryUnited Kingdomen_GB
dc.publisher.countryBritanniafi_FI
dc.publisher.country-codeGB
dc.relation.articlenumbere2018
dc.relation.doi10.1002/leap.2018
dc.relation.ispartofjournalLearned Publishing
dc.relation.issue4
dc.relation.volume38
dc.source.identifierhttps://www.utupub.fi/handle/10024/195189
dc.titleDoes ChatGPT Ignore Article Retractions and Other Reliability Concerns?
dc.year.issued2025

Tiedostot

Näytetään 1 - 1 / 1
Ladataan...
Name:
Learned Publishing - 2025 - Thelwall - Does ChatGPT Ignore Article Retractions and Other Reliability Concerns.pdf
Size:
397.55 KB
Format:
Adobe Portable Document Format