Please use this identifier to cite or link to this item:
https://scholar.dlu.edu.vn/handle/123456789/3554| Title: | ThangDLU at #SMM4H 2024: Encoder-decoder models for classifying text data on social disorders in children and adolescents | Authors: | Tạ, Hoàng Thắng Abu Bakar Siddiqur Rahman Lotfollah Najjar Alexander Gelbukh |
Keywords: | Computer Science - Computation and Language; Computer Science - Computation and Language | Issue Date: | 2024-05-01 | Journal: | Association for Computational Linguistics | Volume: | Proceedings of The 9th Social Media Mining for Health Research and Applications (SMM4H 2024) Workshop and Shared Tasks | Pages: | 1-4 | Abstract: | This paper describes our participation in Task 3 and Task 5 of the #SMM4H (Social Media Mining for Health) 2024 Workshop, explicitly targeting the classification challenges within tweet data. Task 3 is a multi-class classification task centered on tweets discussing the impact of outdoor environments on symptoms of social anxiety. Task 5 involves a binary classification task focusing on tweets reporting medical disorders in children. We applied transfer learning from pre-trained encoder-decoder models such as BART-base and T5-small to identify the labels of a set of given tweets. We also presented some data augmentation methods to see their impact on the model performance. Finally, the systems obtained the best F1 score of 0.627 in Task 3 and the best F1 score of 0.841 in Task 5. |
URI: | https://scholar.dlu.edu.vn/handle/123456789/3554 | URL: | https://aclanthology.org/2024.smm4h-1.1/ | Type: | Bài báo đăng trên KYHT quốc tế (có ISBN) |
| Appears in Collections: | Kỷ yếu hội thảo (Khoa Công nghệ thông tin) |
Files in This Item:
| File | Description | Size | Format | |
|---|---|---|---|---|
| 2024.smm4h-1.1.pdf | 91.84 kB | Adobe PDF | View/Open |
CORE Recommender
Page view(s)
159
Last Week
1
1
Last month
checked on Mar 10, 2026
Download(s)
26
checked on Mar 10, 2026
Google ScholarTM
Check
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.