Skip to content

Hallucination Detection #2

@joelpaulkoch

Description

@joelpaulkoch

After generating the response, we can check if the response contains hallucinations.
In a RAG context, there are several points to consider:

  • does the response answer the question?
  • is it based on the provided context?
  • are there obvious hallucinations?

There are different ways to implement hallucination detection:

  1. Models specifically trained for this task, for instance from vectara
  2. LLMs answering the points above
  3. Other techniques, e.g. SelfCheckGPT

For now I've implemented a simple hallucination detection that makes use of an LLM in ef1763b.

There is the RAG Triad that is used to check for hallucinations.
This could be done by an LLM too, we can ask it to return a score for each of the criteria and classify them as hallucination based on thresholds.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions