Probability is ordinary before it is technical
You already use probability whenever you judge a claim in light of new information. A doctor reads symptoms and test results. A weather forecast changes when new data arrives. A court weighs testimony, motive, physical evidence and alternative explanations. A spam filter asks whether a message fits the pattern of spam or real mail.
Bayesian inference gives that ordinary habit a disciplined form. It says, in effect: begin somewhere, add evidence and update honestly. The method is not magic. It can be abused. But the basic question is sound: does this evidence fit one explanation better than another?
The Signal uses probability as a lamp, not a throne. It helps show whether evidence is moving the question. It does not save, worship or command the soul.
How the weights work
The Signal uses evidence weights to ask whether an item favors one hypothesis over another. A strong piece of evidence should move the map more than a weak one. A modest clue should stay modest. A contextual clue should not pretend to be a proof.
That is why the site keeps caveats visible. Some items are direct. Some are background texture. Some are support-layer evidence. Some are Scripture-to-Scripture coherence. Some are scientific, historical, philosophical or experiential. They do not all do the same work, so they should not all receive the same kind of weight.
The better question is not, "Can a number make this certain?" It is, "Are these weights restrained, inspectable and connected to the actual evidence?" If an item is overstated, it should be corrected. If two items overlap, the dependency controls should keep them from being counted twice as though they were independent.
Stated priorsBayes factorsCaveatsDependency controlsStage context
Why the AI checks matter, and why they do not settle it
As of May 8, 2026, the weighting structure had been tested against major AI systems including Claude, Gemini, DeepSeek and ChatGPT. The point of those tests was not to ask machines for spiritual authority. It was to ask whether independent systems, reading the map, found the weighting structure broadly coherent, restrained and fit for audit.
That kind of agreement is useful. It can reveal whether the map is obviously lopsided, whether the caveats are being ignored or whether the weights look wildly out of proportion. But it is not the foundation. The foundation remains the evidence, the reasoning, the open caveats, the rival explanations and the coherence burden placed on every worldview.
If an AI system says the map is reasonable, that may be worth noticing. If the evidence itself is weak, AI approval cannot make it strong. If the evidence is strong, AI discomfort cannot make it vanish.