版权所有:内蒙古大学图书馆 技术提供:维普资讯• 智图
内蒙古自治区呼和浩特市赛罕区大学西街235号 邮编: 010021
作者机构:Donald Bren School of Information & Computer Sciences University of California IrvineCA United States Luddy School of Informatics Computing & Engineering Indiana University BloomingtonIN United States Instituto Gulbenkian de Ciência Oeiras Portugal Dept. of Systems Science & Industrial Engineering Binghamton University BinghamtonNY United States School of Nursing Indiana University IndianapolisIN United States
出 版 物:《arXiv》 (arXiv)
年 卷 期:2024年
核心收录:
主 题:Eigenvalues and eigenfunctions
摘 要:Objective — To (1) identify health-related terms used on social media posts that do not precisely match the health-related meaning of terms in a biomedical dictionary, (2) decide which terms need to be removed in order to improve the quality of the dictionary in the scope of biomedical text mining tasks, (3) evaluate the effect of removing imprecise terms on such tasks, and (4) discuss how human-centered annotation complements automated annotation in social media mining for biomedical purposes. Materials and Methods — We used a dictionary built from biomedical terminology extracted from various sources such as DrugBank, MedDRA, MedlinePlus, TCMGeneDIT, to tag more than 8 million Instagram posts by users who have mentioned an epilepsy-relevant drug at least once, between 2010 and early 2016. A random sample of 1,771 posts with 2,947 term matches was evaluated by human annotators to identify false-positives. Frequent terms with a high false-positive rate were removed from the dictionary. To study the effect of removing those terms, we constructed knowledge networks using the refined and the original dictionaries and performed an eigenvector-centrality analysis on both networks. OpenAI’s GPT series models were compared against human annotation. Results — Analysis of the estimated false-positive rates of the annotated terms revealed 8 ambiguous terms (plus synonyms) used in Instagram posts, which were removed from the original dictionary. We show that the refined dictionary thus produced leads to a significantly different rank of important terms, as measured by their eigenvector-centrality of the knowledge networks. Furthermore, the most important terms obtained after refinement are of greater medical relevance. In addition, we show that OpenAI’s GPT series models fare worse than human annotators in this task. Discussion — Dictionaries built from traditional clinical terminology are not tailored for social media language and can bias results when used in biomedical infe