Publication Type
Conference Proceeding Article
Version
acceptedVersion
Publication Date
1-2023
Abstract
Social media can be valuable for extracting information about an event or incident on the ground. However, the vast amount of content shared, and the linguistic variants of languages used on social media make it challenging to identify important situational awareness content to aid in decision-making for first responders. In this study, we assess whether pretrained models can be used to address the aforementioned challenges on social media. Various pretrained models, including static word embedding (such as Word2Vec and GloVe) and contextualized word embedding (such as DistilBERT) are studied in detail. According to our findings, a vanilla DistilBERT pretrained language model is insufficient to identify situation awareness information. Fine-tuning by using datasets of various event types and vocabulary extension is essential to adapt a DistilBERT model for real-world situational awareness detection.
Keywords
pretrained models, situational awareness, BERT, fine tuning, vocabulary extension
Discipline
Databases and Information Systems | Social Media
Research Areas
Data Science and Engineering
Publication
2023 56th Hawaii International Conference on System Sciences: Hawaii, January 3-6: Proceedings
First Page
2110
Last Page
2119
Publisher
IEEE Computer Society
City or Country
Hawaii
Citation
LO, Siaw Ling; LEE, Kahhe; and ZHANG, Yuhao.
Is a pretrained model the answer to situational awareness detection on social media?. (2023). 2023 56th Hawaii International Conference on System Sciences: Hawaii, January 3-6: Proceedings. 2110-2119.
Available at: https://ink.library.smu.edu.sg/sis_research/7761
Copyright Owner and License
Authors
Creative Commons License
This work is licensed under a Creative Commons Attribution-NonCommercial-No Derivative Works 4.0 International License.
Additional URL
https://hdl.handle.net/10125/102894