Skip to the content.

GenAI for Good: The Misinformation and Veracity Engine

By Calvin Nguyen, Samantha Lin, Dr. Arsanjani

Our Info

Calvin Nguyen: Linkedin | Github
Samantha Lin: Linkedin | Github

Abstract

Picture of Chenly Insights
The objective of our product, Chenly Insights, is to help users combat misinformation using factual factors and microfactors. As misinformation gets easier to spread due to deep fakes and large social networks, this project empowers users by giving them a veracity (truthness) for the news they receive through a Mesop interface. A user will link and upload a news article of their choice. Then, two different types of AI, generative and predictive, judge the article on appropriate factuality factors. The scores are accumulated and then shown to the user, along with other visualizations related to the article. In addition, these scores, along with metadata related to the article, will be uploaded to a vector database that generative AI can use. This vector database is constantly updated with new information from fact check and news websites. Users can also converse with the AI and ask questions about each article and why a particular score was given. We also found the best prompting types with adjustments to the prompts that gives the highest accuracy.

Dataset

We utilized multiple different data sources to both establish the ground truth and understand factors that indicate misinformation. Generally, the more context and information you give a large language model, it can give more nuanced and thoughtful judgements on a piece of text. The data is as follows:

Flowchart

Our flowchart in Lucidchart showcasing the data journey thorughout this process.

Generative AI Methods

Tool: Google Gemini 1.5 Pro 002 Picture of Genai FC

Where and how we used it:

Prompting techniques and their objectives:

Additional information we provide to Gemini in our prompts:

Predictive AI Methods

Tools: Python, HuggingFace, XGBoosted Decision Tree, Pandas, PyTorch

Built and trained different models for different factuality factors

Website

We created a product, called Chenly Insight, a front-facting website for users to upload articles and grade them based on their perceived veracity. We utilized the predictive AI and generative AI methods listed here. If you would like to see our Figma link for the product and the poster, here it is: Link. We have a couple of philosophies for our website listed below:

Results

Picture results \

Discussion

In our project, we’ve attempted to follow previous work’s suggestions by combining predictive and generative artificial intelligence in detecting misinformation. We successfully developed a simple hybrid system that scores articles’ veracity, and this is a significant step forward in addressing the complex challenge of misinformation in a digital age. This product can be run by anyone using our system.

There are a couple issues with this current process though. The MSE is highly based on the human scores, which is only based on a sample size of 4. Ideally, we would like a team of 20 expert human graders to grade each article on a factuality factor to get an article’s true score.

Our hypothesis that FCoT with VB and SERP would receive the lowest MSE is wrong. Normal prompting with VB and SERP performed the best. This could be due to a couple reasons. Looking at the bar graph, it seems like FCoT struggled to judge our onion article accurately due to its satire, despite doing well for all the other articles. Additionally, we did notice that FCoT did grade each article more critically and provided more reasoning in relation to the microfactors, so it is possible that human graders did not notice that.

Future Direction

  1. Combine generative and predictive AI for a single factuality factor using agents like CrewAI.
  2. Adjust prompting techniques to help LLMs achieve more human-like scoring through longer and more specificized prompting for FCoT.
  3. Expanding more data sources, such as Washington Post Fact Checker, for our vector database.
  4. Make the website live via Google Cloud and a user inputted API keys and run jobs to perform automated scraping for our database.
  5. Implement more factuality factors into our model to grade articles more critically.

Data Ethics

Data:

System:

Acknowledgements

Calvin and Samantha thank Dr.Arsanjani for his mentorship and guidance throughout this project and other groups within section B01 for debugging and advice with coding throughout the project. We would also like to thank the rest of the capstone group (David Sun, Eric Gu, Eric Sun, Jade Zhou, Luran Zhang, and Yiheng Yuan), as they helped with bouncing ideas and keeping our group accountable

References

  1. Jiang, Bohan, Zhen Tan, Ayushi Nirmal, and Huan Liu. 2024. “Disinformation Detection: An Evolving Challenge in the Age of LLMs.” arXiv preprint arXiv:2309.15847. [Link]
  2. P. Qi, W. Hsu, and M. L. Lee, “Sniffer: Multimodal large language model for explainable out-of-context misinformation detection,” ar5iv, https://ar5iv.labs.arxiv.org/html/2403.03170 (accessed Nov. 3, 2024).
  3. Arsanjani. Ali, ”Alternus Vera,” https://alternusvera.wordpress.com/veracity-vectors-for-disinformation-detection (accessed 2024).
  4. Tariq60. 2018. “LIAR-PLUS.” Oct. [Link]