• 0 Posts
  • 5 Comments
Joined 21d ago
cake
Cake day: Jan 03, 2026

help-circle
rss

Huh? That is the literal opposite of what I said. Like, diametrically opposite.

The system summarizes and hashes docs. The model can only answer from those summaries in that mode. There’s no semantic retrieval step.

No, that’s exactly what you wrote.

Now, with this change

SUMM -> human reviews

That would be fixed, but will work only for small KBs, as otherwise the summary would be exhaustive.

Case in point: assume a Person model with 3-7 facts per Person. Assume small 3000 size set of Persons. How would the SUMM of work? Do you expect a human to verify that SUMM? How are you going to converse with your system to get the data from that KB Person set? Because to me that sounds like case C, only works for small KBs.

Again: the proposition is not “the model will never hallucinate.”. It’s “it can’t silently propagate hallucinations without a human explicitly allowing it to, and when it does, you trace it back to source version”.

Fair. Except that you are still left with the original problem of you don’t know WHEN the information is incorrect if you missed it at SUMM time.


The system summarizes and hashes docs. The model can only answer from those summaries in that mode

Oh boy. So hallucination will occur here, and all further retrievals will be deterministically poisoned?


So… Rag with extra steps and rag summarization? What about facts that are not rag retrieval?


A very tailored to llms strengths benchmark calls you a liar.

https://artificialanalysis.ai/articles/gemini-3-flash-everything-you-need-to-know (A month ago the hallucination rate was ~50-70%)


I want to believe you, but that would mean you solved hallucination.

Either:

A) you’re lying

B) you’re wrong

C) KB is very small