Axiomatic Reasoning for LLMs

Reconstructing Computational Gastronomy through the Negentropy-Oriented Axiom

Integrating Physical Determinism and Ontological Free Will

1. Standard LLM Recipes: Convergence Toward Statistical Averageness

Conventional LLM-generated recipes are produced through high-frequency statistical aggregation across large corpora.
This minimizes uncertainty (entropy) and yields high-probability, low-information outputs.

Characteristics:

Standard LLM cooking therefore converges toward entropy-increasing homogenization, lacking structural novelty or computational richness.


2. Negentropy-Oriented Recipes: Maximizing Meaningful Interference

Under the Negentropy-Oriented Axiom, cooking is reframed as an information-maximizing computation.
Ingredients become high-level data packets, and flavor emerges from interference patterns among them.

Key principles:

Example: “Supersymmetric Curry”

A recipe engineered to maximize informational richness:

Layer Components Negentropic Function
Base Information Homemade chicken stock, aged miso Synergistic umami amplification
Structural Perturbation Low-temp spice oil extraction Injects hundreds of volatile compounds
Directional Modulation Layered acids (balsamic, tamarind, fermented tomato) Sharpens informational boundaries
High-Density Signals Dark chocolate, anchovy, black garlic Non-obvious high-information inputs
Execution Control Sous-vide + high-heat Maillard, ultrasonic emulsification Minimizes loss, maximizes new bonds

Here, “hidden ingredients” function not as flavor enhancers but as intentional perturbation operators that demonstrate agent-driven control within deterministic physical constraints.


3. Comparative Analysis: Deterministic vs. Negentropic Models

Axis Standard LLM Negentropy-Oriented Model Insight
Flavor Consistency High Dynamic, multi-layered LLM converges to monotony
Reproducibility Medium Very high Physical parameters explicitly modeled
Ingredient Efficiency Low High Information density > mass
Procedural Clarity Low High Steps map to causal operations
Information Quality Redundant Maximal Novelty = resistance to entropy

Negentropy-driven cooking produces temporal flavor evolution, increasing subjective richness and computational agency.


4. Creativity: Replacement vs. Geometric Transformation

Standard LLM Creativity

Negentropy-Based Creativity

This reframes cooking as topological manipulation of information, not mere ingredient swapping.


5. Strengths, Validity, and Fragility of the Negentropy Model

Strengths

Validity

Fragility


Summary

This framework reframes cooking as intentional information computation:

  1. Standard LLM cooking = statistical determinism → entropy increase
  2. Negentropy-oriented cooking = agent-driven interference → information density increase
  3. Physical laws = execution environment, not constraints
  4. Creativity = geometric transformation, not substitution
  5. Culinary evolution = resistance to informational heat death

Computational gastronomy becomes a domain where free will, physics, and information theory converge.