Tackling the “Truthiness” Problem: Ensuring Reliability in Large Language Models

Guido Maciocci
5 min readJan 4, 2024

--

Image by the author

The stellar rise of large language models (LLMs) has sparked excitement about their potential to transform how we interact with technology and the way we work. Whether it’s writing code, generating documents, summarizing reports, or analyzing data, LLMs display an impressive ability to generate remarkably human-like text.

However, their uncanny fluency can sometimes be misleading. LLMs have a penchant for “hallucinating” — making up convincing-sounding statements without any factual basis. This tendency, described as their “truthiness problem”, poses a major roadblock for the safe and ethical deployment of LLMs, especially in sensitive domains like healthcare and finance.

LLMs are already influencing critical business decision-making around the world. Designing systems that enable trust, transparency, and verifiability of generated outputs will be fundamental to the development of reliable, ethical, and responsible AI applications.

This article explores the importance of both technical approaches — including prompt engineering, knowledge retrieval, and self-refinement — and user experience design, in mitigating hallucinations and creating applications that enable trustworthy relationships between humans and LLMs.

Why Verifiability Matters

Unlike most traditional AI systems designed for narrow tasks, LLMs are exposed to vast swathes of online text during training. This allows them to extrapolate information and mimic human language, but also means they easily pick up on biases in the data and have a penchant for generating convincing falsehoods. Small errors could snowball into catastrophic outcomes when leveraging LLMs’ generation capabilities for high-stakes scenarios like legal consultation or investment recommendations.

To responsibly deploy LLMs in enterprise settings, it’s vital to ensure their outputs are verifiable — i.e. the lineage of statements can be traced back to credible sources. This allows end-users to “trust but verify” suggestions. UX design plays a key role here by incorporating interactive features that enable transparency, such as evidence linking and explanations.

Platforms like Perplexity.ai, Scite.ai, Cohere, and ChatGPT are pioneering novel user experiences that include citations of external references and provide users the ability to inspect the reliability of statements, addressing the black-box nature of LLMs. Such innovations will be crucial as enterprises increasingly look to harness the power of LLMs to augment human capabilities and enhance decision-making.

Emerging Strategies for Tackling Hallucinations

A recently published survey paper introduces an insightful taxonomy categorizing the diverse techniques researchers have introduced to address the hallucination problem. These range from refinements applied after initial text generation to novel model architectures designed from scratch to be more faithful.

Post-Generation Refinement

Methods like Retrofit Attribution using Research and Revision (RARR) and High Entropy Word Spotting optimize LLM outputs after the initial generation phase. RARR mimics fact-checking workflows — retrieving evidence to verify claims in the generated text and editing to align with credible sources. High Entropy Word Spotting identifies speculative terms in LLM outputs that lack factual grounding and replaces them with less ambiguous alternatives using a secondary LLM.

Self-Refinement via Feedback

Other techniques employ a feedback loop — checking for flaws like contradictory statements in the LLM’s initial output and having it refine its response accordingly. For instance, the ChatProtect method formulates a 3-step pipeline to pinpoint contradictory statements and mitigate their occurrence through targeted prompting.

Prompt Engineering

Carefully engineering the instructions and examples provided to prime LLMs is emerging as an effective strategy. Methods like UPRISE and SynTra optimize model prompts to reduce hallucinations for specific downstream tasks like dialogue and summarization.

Supervised Fine-Tuning

Adjusting model weights by fine-tuning on labeled datasets is another avenue, with techniques like Knowledge Injection directly training models on relevant domain knowledge. HAR generates counterfactual datasets by hallucinating incorrect details about factual topics, helping models better distinguish truth from fiction.

Novel Model Architectures

Some studies design new models from scratch with mechanisms to enhance faithfulness. This includes context-aware decoding strategies like CAD that override contradictions between an LLM’s prior knowledge and given input, as well as loss functions that penalize hallucinations.

Designing Trust: The Role of User Experience Design

The way we consume AI-generated outputs and interact with AI models plays a crucial role in shaping the “truthiness” of LLMs and fostering trustworthiness in their outputs. A well-designed UX can make the difference between a user blindly accepting AI-generated content and a user who is equipped to critically assess and validate that information.

Transparency is key. UX should be designed to provide users with clear insights into how the LLM operates, its limitations, and the source of its information. For instance, when an LLM provides an answer, the UX could include a feature that allows users to see a “confidence score” or a summary of the data sources used. Such features empower users to gauge the reliability of the information presented.

Another aspect is the integration of user feedback mechanisms. Allowing users to flag uncertain or incorrect responses helps in two ways: it provides data to improve the model, and it also gives users a sense of control and participation in the AI ecosystem. This feedback loop can be a powerful tool in building trust, as users see their input valued and acted upon.

Providing options for customization can also enhance trust. Users having the ability to set preferences for how they receive information, like choosing between detailed explanations or concise summaries, allows for a personalized experience. This personalization respects the user’s individual needs and learning styles, which is crucial in building a trusting relationship between the user and the AI.

Creative UX designs enabling transparency and a “trust but verify” approach will be just as crucial as model architecture and prompting advances — allowing end-users to seamlessly validate the lineage of statements made by LLMs.

Trustworthy LLMs

While each technique makes unique contributions, long-term solutions likely entail an amalgamation of strategies tailored to specific use cases. Reducing reliance on labeled data through unsupervised and weakly supervised techniques could improve scalability. Architectures that synthesize self-correction abilities with grounded knowledge retrieval offer promise. Coming up with universally effective strategies requires continued collaboration between researchers, developers, and domain experts.

The UX of LLMs is not just about aesthetics or ease of use. It’s a fundamental component in building a trustworthy relationship between the user and the AI. By prioritizing transparency, feedback, accessibility, customization, and continuous improvement, UX designers can create environments where users are not only satisfied with the AI’s performance but are also empowered to “trust but verify”.

As LLMs continue maturing in both capabilities and safety, their position in business ecosystems is guaranteed to expand dramatically. Verifiable LLMs that enrich human intelligence while rigorously validating information have the potential to transform workflows across sectors like finance, law, and healthcare. With diligent efforts to enhance model fidelity and thoughtful UX choices to enable user transparency, we can harness the strengths of LLMs for collaborative decision-making while mitigating their pitfalls.

If you’d like to find out more about how you can leverage Large Language Models to help your organization work smarter, feel free to reach out to me at at AecFoundry or add me on LinkedIn

--

--

Guido Maciocci

Technology, Strategy, and Product | AEC Industry Founder @ AecFoundry