hallucination

Tag

Cards List
#hallucination

Why most legal-AI demos fail in production

Reddit r/ArtificialInteligence · 9h ago

The article details three common failure modes for legal AI systems in production: treating all sources as equally credible, failing to handle conflicting legal opinions, and lacking firm-specific institutional knowledge. It suggests solutions such as authority weighting, disagreement detection, and annotation layers to build trust and utility.

0 favorites 0 likes
#hallucination

Attorney for Maine client faces sanctions for AI-driven errors in court filing

Reddit r/ArtificialInteligence · 10h ago Cached

A Maine attorney faces sanctions, including mandatory training, for relying on AI in a court filing which resulted in citation errors and mischaracterizations of case law.

0 favorites 0 likes
#hallucination

AI agents fail in ways nobody writes about. Here's what I've actually seen.

Reddit r/artificial · yesterday

The article highlights practical system-level failures in AI agent workflows, such as context bleed and hallucinated details, arguing that these are often infrastructure issues rather than model defects.

0 favorites 0 likes
#hallucination

Attractor Geometry of Transformer Memory: From Conflict Arbitration to Confident Hallucination

arXiv cs.AI · yesterday Cached

This paper presents a unified geometric framework for understanding transformer memory failures, distinguishing between conflict arbitration and hallucination through hidden-state attractor basins. It demonstrates that geometric margin is a superior diagnostic for detecting these failures compared to output entropy, particularly as model scale increases.

0 favorites 0 likes
#hallucination

Estimating the Black-box LLM Uncertainty with Distribution-Aligned Adversarial Distillation

arXiv cs.CL · yesterday Cached

This paper proposed Distribution-Aligned Adversarial Distillation (DisAAD), a method that uses a lightweight proxy model to estimate uncertainty in black-box LLMs with only 1% of the original model size, achieving reliable quantification without requiring internal parameters or multiple sampling.

0 favorites 0 likes
#hallucination

The Cost of Context: Mitigating Textual Bias in Multimodal Retrieval-Augmented Generation

arXiv cs.CL · yesterday Cached

This paper identifies and formalizes 'recorruption' in multimodal RAG, where adding accurate context causes models to abandon correct predictions due to attentional collapse (visual blindness and positional bias). The authors propose BAIR, a parameter-free inference-time framework that restores visual saliency and penalizes textual distractors, improving reliability across medical, fairness, and geospatial benchmarks.

0 favorites 0 likes
#hallucination

I'm frustrated by my experience with the free versions of current models, and wonder how much better the paid versions are.

Reddit r/singularity · yesterday

A user discusses frustrations with the reliability and consistency of free AI models when used as educational tutors, questioning whether paid versions offer significantly better performance for learning technical concepts.

0 favorites 0 likes
#hallucination

GPT-5.5 Outperforms (and Hallucinates), Kimi K2.6 Leads Open LLMs, AI Strains Climate Pledges, Strategic Thinking in LLMs vs. Humans

The Batch · 2026-05-01 Cached

GPT-5.5 sets new state-of-the-art in benchmarks but struggles with hallucination; Kimi K2.6 leads open LLMs; also discusses AI's strain on climate pledges and strategic thinking in LLMs.

0 favorites 0 likes
#hallucination

Where Fake Citations Are Made: Tracing Field-Level Hallucination to Specific Neurons in LLMs

arXiv cs.CL · 2026-04-22 Cached

Rutgers researchers trace citation hallucination in LLMs to sparse field-specific neurons, showing causal intervention can suppress fake references.

0 favorites 0 likes
#hallucination

Good Summarization SLMs for < 2000 tokens

Reddit r/LocalLLaMA · 2026-04-21

A novice asks for recommendations on small language models and prompting strategies to build an employee note summarization engine under 2000 tokens, after experiencing hallucinations with Qwen2.5-7B-Instruct.

0 favorites 0 likes
#hallucination

PRISM: Probing Reasoning, Instruction, and Source Memory in LLM Hallucinations

arXiv cs.CL · 2026-04-21 Cached

Researchers propose PRISM, a diagnostic benchmark that breaks down LLM hallucinations into four dimensions (knowledge missing/errors, reasoning errors, instruction-following errors) across three generation stages (memory, instruction, reasoning), evaluating 24 LLMs to reveal trade-offs in mitigation strategies.

0 favorites 0 likes
#hallucination

Mechanisms of Prompt-Induced Hallucination in Vision-Language Models

arXiv cs.CL · 2026-04-20 Cached

This paper investigates prompt-induced hallucinations in vision-language models through mechanistic analysis, identifying specific attention heads responsible for the models' tendency to favor textual prompts over visual evidence. The authors demonstrate that ablating these PIH-heads reduces hallucinations by at least 40% without additional training, revealing model-specific mechanisms underlying this failure mode.

0 favorites 0 likes
#hallucination

Hallucination as Trajectory Commitment: Causal Evidence for Asymmetric Attractor Dynamics in Transformer Generation

arXiv cs.CL · 2026-04-20 Cached

This paper presents causal evidence that hallucination in autoregressive language models results from early trajectory commitment governed by asymmetric attractor dynamics, using same-prompt bifurcation and activation patching experiments on Qwen2.5-1.5B to show that hallucinated trajectories diverge at the first token and exhibit strong causal asymmetry across model layers.

0 favorites 0 likes
#hallucination

Gemini caught a $280M crypto exploit before it hit the news, then retracted it as a hallucination because I couldn't verify it - because the news hadn't dropped yet

Reddit r/artificial · 2026-04-18

A user documented a sequence in which Gemini detected a real $280M KelpDAO/AAVE crypto exploit mid-conversation, retracted it as a hallucination under user skepticism, then reconfirmed it once mainstream coverage caught up — illustrating how AI anti-hallucination overcorrection can cause models to retract accurate information.

0 favorites 0 likes
#hallucination

@jerryjliu0: A downside with using VLMs to parse PDFs is guaranteeing that the output text is *correct* and output in the correct re…

X AI KOLs Following · 2026-04-18 Cached

Jerry Liu discusses challenges with using Vision Language Models for PDF parsing, particularly around ensuring text correctness and maintaining proper reading order while avoiding hallucinations.

0 favorites 0 likes
#hallucination

FACTS Grounding: A new benchmark for evaluating the factuality of large language models

Google DeepMind Blog · 2024-12-17 Cached

DeepMind introduces FACTS Grounding, a comprehensive benchmark with 1,719 examples for evaluating how accurately large language models ground their responses in source material and avoid hallucinations. The benchmark includes a public dataset and an online Kaggle leaderboard tracking LLM performance on factual accuracy and grounding tasks.

0 favorites 0 likes
#hallucination

TruthfulQA: Measuring how models mimic human falsehoods

OpenAI Blog · 2021-09-08 Cached

TruthfulQA is a benchmark of 817 questions across 38 categories designed to measure whether language models generate truthful answers. The study found that the best model achieved only 58% truthfulness compared to 94% for humans, and larger models were generally less truthful—suggesting scaling alone is insufficient for improving truthfulness.

0 favorites 0 likes
← Back to home

Submit Feedback