Skip to content

OriginalityAI/Fact-Checker-LLM-Study

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

14 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Overview

A dataset of 5000 claims from 5 Large Language Models was processed by the Originality.ai Fact Checker.

Data

1000 prompts from 10 broad topic categories was passed through the GPT-3.5, GPT-4, LLAMA-7B, LLAMA-13B and LLAMA-70B to generate claims. These claims were then processed through Originality.ai's Fact Checker which generated scores of True, False or NAN in situations where the LLM refused to generate a claim.

Key Findings

  • Average Model Accuracy across topics: 67.9% (LLAMA-13) to 76.9% (GPT-4)
  • Average topic accuracy across models: Health (80.5%) to News (64.4%).
  • 100% Prompt Confidence for LLAMA models. and 97% Confidence for the GPT models.

plot

Disclaimer

The Originality.ai Fact Checking tool is an aid to support editors to more efficiently fact check claims. and it is up to the end-user to interpret the results appropriately. It will sometimes provide inaccurate responses. This can include the data used in this study.

About

We used our Fact Checker tool to test the accuracy of claims generated by 5 Large Language models: GPT-3.5, GPT-4, LLAMA-7B, LLAMA-13B and LLAMA-70B.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors