SLMs (Small Language Models) for Misinformation Detection

Abstract
People are overwhelmed by a huge amount of news they receive every day. Many times, they may relay the news to their family members or friends if they deem the news is relevant to the family members or friends, but not all the news they receive is true. According to a study, 38% of Americans have the experience of sharing fake news with others, and people usually trust the news more if it is from the persons they know. Various machine learning methods have been used to identify fake news, and each has its pros and cons. Currently, generative AI and LLMs (large language models) are trendy and very popular: However, the above two methods require a tremedous amount of resources like thousands of GPUs and trills of parameters. SLMs (Small Language Models) are created to solve the resource problems. This research is to apply SLMs to misinformation processing and hope better results will be achieved. Two applications of SLMs related to misinformation management will be considered are LLMs are based on NLP (natural language processing) and misinformation is a kind of text. If the LLMs are well trained, they should be able to tell whether the information is true or false. On the other hand, misinformation could be synthesized by LLMs. How to identify the misinformation synthesized by LLMs could be another research topic.

Keywords
Security, misinformation, misinformation identification, LLMs, large language models, natural language processing

References
Combating misinformation in the age of LLMs: Opportunities and Challenges
Creating a large language model from scratch: A beginner’s guide

Conference