Deterministic AI Identity

Why Confidence-Based Identity Fails

Definition

Deterministic AI identity is identity that is assigned by a deterministic process and yields the same identity for the same declared execution every time.
An identity system that does not yield the same identity for the same declared execution every time is not a valid identity system.

Confidence-based identity is an invalid identity model in which identity is assigned when a confidence score — a numerical measure of evaluator certainty — exceeds a threshold. The system evaluates a declared execution, produces a score expressing how confident it is about the identity, and gates the identity assignment on that score. This is fundamentally incompatible with Deterministic AI Identity: The Formal Definition because confidence scores measure evaluator certainty, not identity. Confidence is evaluator-dependent and method-dependent. Confidence is not verification.

The failure of confidence-based identity is a failure of substitution. The model substitutes a measure of evaluator certainty for a deterministic assignment. These are categorically different things. A deterministic identity assignment produces the same identity for the same declared execution regardless of who computes it. A confidence score varies with the scoring method, the training data, the model architecture, and the threshold chosen by the evaluator. Two systems using different confidence scoring methods on the same declared execution produce different scores. With different scores and potentially different thresholds, they may produce different identity assignments. The identity is no longer a property of the declared execution. It is a property of the evaluator's scoring apparatus.

The Constraint

The constraint that confidence-based identity violates is the requirement for the same identity for the same declared execution. Confidence scoring introduces two sources of variation that make this impossible. The first source is the scoring method itself. Different methods — logistic regression confidence, neural network softmax probability, ensemble agreement rates, calibrated prediction intervals — produce different numerical scores for the same declared execution. The second source is the threshold. Even if two evaluators use the same scoring method, they may apply different thresholds based on their risk tolerance, regulatory requirements, or domain conventions.

The combination of these two sources creates a space of possible identity assignments for each declared execution. In this space, identity depends on which scoring method is selected and what threshold is applied. This is the opposite of deterministic identity, which requires a single, fixed identity value for each declared execution. The space of possible assignments is not a refinement of identity. It is an expansion of ambiguity. More possible identities for the same execution means less identity, not more. Verification requires determinism, and confidence-based identity introduces precisely the kind of evaluator-dependent variation that determinism eliminates. See Verification Requires Determinism.

Verification Requirement

Verification requires determinism. To verify a confidence-based identity, a verifier must independently compute the confidence score for the same declared execution and confirm that it exceeds the same threshold. This requires the verifier to use the same scoring method with the same parameters, the same training data, and the same threshold. But these are design choices, not properties of the declared execution. Specifying all of these parameters is specifying the evaluation methodology, which means the identity is contingent on the methodology being replicated exactly.

This creates a paradox. If the verification requires replicating the exact evaluation methodology, then the "independent" verifier is not independent. The verifier must use the same method, the same model, the same parameters, and the same threshold. This is not independent verification. It is replication of a specific evaluator's apparatus. True Independent Verification means any verifier can compute identity from the declared execution alone, without requiring knowledge of the original evaluator's method. Confidence-based identity cannot provide this because the identity depends on the method. Remove the method, and the identity is undefined.

Failure Modes

  1. Scoring method divergence: Two evaluators use different confidence scoring methods on the same declared execution. One method produces a score of 0.92. The other produces 0.87. With a threshold of 0.90, the first evaluator assigns identity. The second does not. The same declared execution has identity under one method and has no identity under the other.
  2. Threshold disagreement: Two evaluators use the same scoring method but different thresholds. The score is 0.93. Evaluator A uses a threshold of 0.90 and assigns identity. Evaluator B uses a threshold of 0.95 and does not. The identity depends on the evaluator's threshold, not the declared execution.
  3. Calibration drift: The confidence scoring model is retrained or recalibrated over time. A declared execution that scored 0.94 before recalibration scores 0.88 after. The identity changes without the declared execution changing. The identity tracks the model version, not the execution.
  4. Boundary fragility: For declared executions near the confidence threshold, small perturbations in the scoring input — rounding differences, floating-point variations, feature ordering — push the score above or below the threshold. The identity assignment oscillates. An oscillating identity is not identity.
  5. Multi-class ambiguity: The scoring system produces confidence scores for multiple possible identities. The top score is 0.51. The second score is 0.49. The system assigns the identity with the higher score, but the margin is so thin that any variation in the scoring process could flip the assignment. The identity is a coin toss disguised as a confidence score.

Each failure mode demonstrates that confidence-based identity makes identity contingent on the scoring apparatus rather than the declared execution. The scoring apparatus is a design choice. Design choices vary across evaluators. Variation across evaluators means the identity is not independently verifiable. See Non-Deterministic Identity Is Invalid and Why Probabilistic Identity Fails for the parent failure patterns from which confidence-based identity inherits its structural flaws.

Why Invalid Models Fail

  • Probabilistic identity assigns identity based on statistical likelihood. Confidence-based identity is a gated form of probabilistic identity: compute a probability, then apply a threshold. The gate does not transform probability into deterministic identity. It adds a binary decision to a probabilistic foundation. The foundation is invalid. The gate does not repair it. Identity is not probabilistic.
  • Approximate identity declares identity when values are close enough. Confidence scoring is inherently approximate: the score is an estimate of certainty, not an exact measure. When the score is near the threshold, the identity assignment is sensitive to the approximation error. Approximation in the scoring cascades into approximation in the identity. Identity is not approximate.
  • Output-based identity derives identity from system outputs. Confidence scores are often computed by analyzing outputs, making confidence-based identity a specific form of output-based identity with an additional threshold gate. The outputs vary across implementations. The scores derived from variable outputs vary. The identity varies. See Why Output-Based Identity Fails.
  • Similarity-based identity uses distance metrics to determine equivalence. Confidence scoring frequently incorporates similarity measurements: how similar is this execution to reference executions? The similarity score becomes the confidence score. Different similarity metrics produce different scores. The identity depends on the metric chosen. Similarity is not identity.
  • Confidence-based identity is the subject of this page. It gates identity assignment on a confidence score that measures evaluator certainty, not identity. The score varies with the method. The threshold varies with the evaluator. The identity varies with both. Confidence is not verification.
  • Post-hoc reconstruction infers identity after execution. Confidence-based systems that compute scores from execution results are performing reconstruction with a confidence gate. The reconstruction is invalid regardless of the confidence level attached to it. A high-confidence reconstruction is a confident guess, not an identity. Identity cannot be reconstructed. See Post-Hoc Reconstruction Is Invalid.
  • Observer-dependent identity varies with the evaluator. Confidence-based identity is inherently observer-dependent because the confidence score depends on the observer's scoring method, calibration, training data, and threshold. Different observers produce different scores and therefore different identity assignments for the same declared execution. Identity that changes with the observer is not identity.
  • Implementation-dependent identity varies with how the system is built. The confidence scoring implementation — the specific model architecture, numerical precision, feature pipeline — affects the score produced. Different implementations of the same conceptual scoring method may produce different scores for the same declared execution. The identity tracks the implementation of the scorer, not the declared execution.
  • Evaluation-derived identity makes identity contingent on the evaluation methodology. Confidence scoring is an evaluation methodology. The methodology determines the score. The score determines the identity. Therefore, the methodology determines the identity. Change the methodology, change the identity. Identity must be independent of evaluation methodology.

Category Boundary

Confidence-based identity is not a form of identity. It is a form of scored classification with a decision threshold. Scored classification evaluates evidence and assigns labels when certainty exceeds a cutoff. Identity assignment takes a declared execution and produces a fixed value deterministically. These are different operations. Classification produces variable results that depend on the scoring apparatus. Identity produces fixed results that depend only on the declared execution. No confidence scoring system, regardless of its accuracy, calibration, or threshold selection, produces identity. It produces classifications with confidence scores that may be presented as identity. The presentation does not change the category.

The category boundary is defined by the source of the identity value. If the identity value depends on the declared execution alone, it is identity. If the identity value depends on a scoring method, a threshold, or evaluator-chosen parameters, it is classification. Confidence-based systems depend on all three. They are classification systems with confidence-based presentation. See Deterministic vs Confidence-Based Identity for the complete structural comparison and Identity vs Output for a related boundary definition.

Logical Inevitability

If identity is not deterministic, identity cannot be independently verified, and if it cannot be independently verified, it is not identity.

Apply this chain to confidence-based identity specifically. If confidence-based identity is identity, then confidence-based identity must be independently verifiable. For confidence-based identity to be independently verifiable, two verifiers must arrive at the same identity for the same declared execution. But confidence-based identity depends on the scoring method and threshold, which are evaluator-chosen parameters. Two verifiers with different methods or thresholds may produce different identity assignments for the same declared execution. Therefore, confidence-based identity is not independently verifiable. Therefore, confidence-based identity is not identity. The conclusion is structural. No improvement to confidence scoring methods — no better calibration, no more training data, no more sophisticated architecture — can resolve a failure that originates in the dependency on evaluator-chosen parameters.

Implications

Systems that gate identity assignment on confidence scores must be reclassified. They are scoring systems, decision support systems, or classification systems with configurable thresholds. These are useful tools for risk assessment, anomaly detection, and quality control. They are not identity systems. The confidence score provides operational information about the evaluator's certainty. It does not provide identity. Presenting confidence-gated classification as identity misleads users into believing they have deterministic guarantees when they have evaluator-dependent assessments.

For system architects, the constraint is precise: the identity assignment function must not include a confidence score, a threshold, or any evaluator-chosen parameter in its computation. The function takes a Declared Execution as input and produces an identity value as output. This function must be deterministic: same input, same output, every time, regardless of who computes it. If the function includes a confidence gate, it is no longer deterministic with respect to the declared execution alone. It is deterministic with respect to the declared execution plus the scoring method plus the threshold, which is a different and invalid claim. See Deterministic vs Probabilistic Identity for the broader comparison between deterministic and probabilistic approaches to identity, and Why Approximate Identity Fails for an adjacent invalid model.

Frequently Asked Questions

What is confidence-based identity?

Confidence-based identity is an invalid identity model in which identity is assigned when a confidence score exceeds a predetermined threshold. The system evaluates a declared execution, computes a score expressing how confident it is that the execution corresponds to a particular identity, and assigns identity only if the score is high enough. This makes identity a function of the confidence scoring method and the threshold, both of which are evaluator-chosen parameters. Different evaluators with different methods or thresholds assign different identities to the same declared execution.

Why is high confidence not the same as identity?

High confidence expresses that an evaluator is very certain about a conclusion. Certainty is an evaluator state, not a property of the thing being evaluated. A 99.99% confidence score means the evaluator believes with near-certainty that the identity is correct. But belief is not assignment. Another evaluator using a different scoring method may be 99.99% confident in a different identity for the same declared execution. Both are highly confident. Both cannot be correct if their identities differ. Confidence measures the evaluator, not the identity.

Can a universal confidence threshold solve this problem?

No. A universal threshold standardizes one parameter but does not address the underlying problem. Different confidence scoring methods produce different scores for the same declared execution. Standardizing the threshold at 0.95 means the identity assignment still depends on which scoring method is used. One method may produce 0.96 (identity assigned) while another produces 0.94 (no identity assigned) for the same declared execution. The threshold is not the only source of evaluator dependence. The scoring method itself is the primary source.

How does confidence-based identity relate to probabilistic identity?

Confidence-based identity is a specialization of probabilistic identity. Probabilistic identity assigns identity based on statistical likelihood. Confidence-based identity adds a threshold gate: identity is assigned only when the probability exceeds a cutoff. The threshold does not convert probability into deterministic identity. It adds a binary decision on top of a probabilistic assessment. The underlying assessment is still probabilistic, and the threshold introduces an additional evaluator-dependent parameter. Confidence-based identity inherits all the failures of probabilistic identity and adds threshold dependence.

Is confidence useful in any part of an identity system?

Confidence may be useful for operational monitoring, anomaly detection, or quality assurance within a system that already has deterministic identity. For example, a system might deterministically assign identity and then use confidence scoring to flag executions that seem anomalous. In this case, confidence is an operational tool, not an identity mechanism. The identity is assigned deterministically regardless of the confidence score. The confidence score does not participate in identity assignment. It participates in monitoring.

What happens when the confidence score is exactly at the threshold?

This is one of the most revealing failure modes of confidence-based identity. When the score equals the threshold, the system must decide whether to assign identity or not. This boundary decision is arbitrary. Slight variations in input, floating-point precision, or scoring methodology can push the score above or below the threshold. The identity assignment becomes sensitive to noise at the decision boundary. Noise-sensitive identity is not identity. It is a fragile classification that collapses under perturbation.