咨询与建议

看过本文的还看了

相关文献

该作者的其他文献

文献详情 >Refinement of an Epilepsy Dict... 收藏
arXiv

Refinement of an Epilepsy Dictionary through Human Annotation of Health-related posts on Instagram

作     者:Min, Aehong Wang, Xuan Correia, Rion Brattig Rozum, Jordan Miller, Wendy R. Rocha, Luis M. 

作者机构:Donald Bren School of Information & Computer Sciences University of California IrvineCA United States Luddy School of Informatics Computing & Engineering Indiana University BloomingtonIN United States Instituto Gulbenkian de Ciência Oeiras Portugal Dept. of Systems Science & Industrial Engineering Binghamton University BinghamtonNY United States School of Nursing Indiana University IndianapolisIN United States 

出 版 物:《arXiv》 (arXiv)

年 卷 期:2024年

核心收录:

主  题:Eigenvalues and eigenfunctions 

摘      要:Objective — To (1) identify health-related terms used on social media posts that do not precisely match the health-related meaning of terms in a biomedical dictionary, (2) decide which terms need to be removed in order to improve the quality of the dictionary in the scope of biomedical text mining tasks, (3) evaluate the effect of removing imprecise terms on such tasks, and (4) discuss how human-centered annotation complements automated annotation in social media mining for biomedical purposes. Materials and Methods — We used a dictionary built from biomedical terminology extracted from various sources such as DrugBank, MedDRA, MedlinePlus, TCMGeneDIT, to tag more than 8 million Instagram posts by users who have mentioned an epilepsy-relevant drug at least once, between 2010 and early 2016. A random sample of 1,771 posts with 2,947 term matches was evaluated by human annotators to identify false-positives. Frequent terms with a high false-positive rate were removed from the dictionary. To study the effect of removing those terms, we constructed knowledge networks using the refined and the original dictionaries and performed an eigenvector-centrality analysis on both networks. OpenAI’s GPT series models were compared against human annotation. Results — Analysis of the estimated false-positive rates of the annotated terms revealed 8 ambiguous terms (plus synonyms) used in Instagram posts, which were removed from the original dictionary. We show that the refined dictionary thus produced leads to a significantly different rank of important terms, as measured by their eigenvector-centrality of the knowledge networks. Furthermore, the most important terms obtained after refinement are of greater medical relevance. In addition, we show that OpenAI’s GPT series models fare worse than human annotators in this task. Discussion — Dictionaries built from traditional clinical terminology are not tailored for social media language and can bias results when used in biomedical infe

读者评论 与其他读者分享你的观点

用户名:未登录
我的评分