me

Veracity AI

How the AI works

Large Language Models (LLMs) are a powerful kind of instruction-ready AI chatbot. They are good at summarizing text like subject-matter experts would. Unlike experts, they can sometimes get basic things wrong in any number of ways. That's why asking them to reason on their own about how likely it is that a statement is true or false is not a reliable approach to using LLMs for fact-checking.

Veri-fact.ai takes a more reliable approach based on “Retrieval-Augmented Generation” or RAG for short. Here, this means having the LLM base its response on relevant text from reliable sources on the internet. Veri-fact.ai does this by:

  1. Instructing the LLM to first come up with text that when pasted in the search bar of a search engine like Google Search will find content relevant to the statement being assessed. The search then returns a list of websites.
  2. Veri-fact.ai then extracts a main snippet of text from each website as well as a source credibility score obtained from a database we use.
  3. The LLM is then instructed to summarize all the sources’ text and scores with the purpose of answering how true is the original statement. The LLM is also instructed to provide a corresponding reliability score between 0 and 100% from very likely false to very likely true. The score is pulled out of the LLM directly--it is not calculated from an equation.

There are a few minor additional steps to this process, e.g.:

That's all there is to it. Now you know! See our User Guidelines for more details on how to use the app.

An academic paper of an earlier version of the approach that goes into more technical detail is:

Jacob-Junqi Tian, Hao Yu, Yury Orlovskiy, Tyler Vergho, Mauricio Rivera, Mayank Goel, Zachary Yang, Jean-François Godbout, Reihaneh Rabbany, & Kellin Pelrine (2024). Web Retrieval Agents for Evidence-Based Misinformation Detection. In First Conference on Language Modeling. [reference]

Some additional notes: