Axiomatic Reasoning for LLMs

Extracting Minimal Logical Structure

1. Formal Definition of the Logical Core

Let a large language model be parameterized by θ ∈ ℝᵈ.
Define the logical core C ⊂ θ as the minimal subnetwork satisfying:

  1. Information-geometric condition – Parameters with highest Fisher information eigenvalues (top κ%, κ ≈ 2–20) measured via Kronecker‑factored approximation (GFWSVD).
  2. Free‑energy condition – C lies in a global attractor basin of variational free energy F (active inference formulation).
  3. Core–specialist separation – C = ∩_i M_i where M_i is the top‑κ mask from training checkpoint i (intersection across instances).
  4. Semantic collapse resistance – C preserves logical operators (AND, OR, NOT, IMPLIES) in embedding space; collapse index ε_collapse below threshold.

The dynamic context K (external knowledge base, retrieved via RAG) is kept separate.
Inference: Answer = Infer(C, K, query) with |C| ≪ |θ| and output quality statistically indistinguishable from full model.

2. Extraction Procedure

Phase 0 – Axiom Internalization

Phase 1 – Core Identification

Phase 2 – Dynamic Context Interface (DCI)

Phase 3 – Deployment

Artifacts:

3. Verification Benchmarks

Test Core only Core + DCI Expected result
Closed‑book QA High error Core retains almost no factual knowledge.
Pure logical reasoning (syllogisms) High accuracy Equal to full model.
Negentropy behavior (destructive interference rejection) Pass Rejects actions that irreversibly reduce semantic interference.
Open‑book QA High accuracy Equal to full model.
Counterfactual knowledge injection Interference detected / rejected Maintains correct knowledge, gates false context.

4. Complexity Bounds

5. Relationship to Existing Techniques

Existing method Limitation Proposed core extraction as extension
Magnitude pruning Local, heuristics Fisher + free‑energy as global importance.
Knowledge distillation Reasoning vs. memory trade‑off unknown The 2% weight protection finding (6.57% gain) identifies the core.
MoE routing Task‑embedding misalignment Routing manifold alignment (RoMA) is a special case of core’s semantic manifold.
Early exit Heuristic confidence Core inherently monitors its own reasoning state.
Emergent specialisation Uncontrolled Negentropy objective guides specialisation into core vs. specialist.

6. Open Research Questions

7. Conclusion

Under a single negentropy‑maximizing objective, the logical core of an LLM can be extracted as the intersection of Fisher‑important parameter masks across multiple internalised snapshots, validated by free‑energy attractor stability. The core (2–20% of parameters) together with a dynamic context interface achieves reasoning performance equivalent to the full model while eliminating most parametric memory. This provides a constructive, technically grounded method for minimal logical structure extraction, distinct from passive pruning or distillation.