E𒂗N𒈾K𒆠A𒀀I𒄿D𒁺U𒌑

𒂗𒆖𒁲

Ex Luto Ad Astra. From Mud to the Stars.

The dominant architectures in modern machine learning were not derived from any physical or mathematical principle. They were found by experiment and scaled. ENKAIDU develops a formal foundation rooted in entropy, energy, and wave dynamics. Model classes, inference procedures, and architectures follow from it as consequences of the theory itself.

TYPE Frontier Intelligence Lab METHOD Entropy and energy methods. Wave dynamics. Statistical mechanics. Computational experiment.
001 // In Plain English

What is happening here?

Most of modern machine learning was found by experiment, not derived from theory. ENKAIDU is building the missing layer: a framework grounded in physics and mathematics where computation, learning, and inference are derived, not engineered.

What we are

A frontier intelligence lab

A research institution that derives computational architectures from mathematical and physical theory, then builds them. Derivation and engineering are one continuous programme.

What we study

The physics beneath learning

Entropy, energy, and wave dynamics govern a class of computation that has not yet been exploited. We study the information theory, statistical mechanics, and mathematics from which it can be derived.

Where we are aiming

A new computational foundation

The objective is to establish a theoretical and engineering base for a different paradigm, and build the systems and tools that follow from it.

002 // Systems Doctrine

How the laboratory operates

The research programme defines the questions. The doctrine defines the methodological standards: what assumptions the lab will not make, what evidence it requires, and how results are expected to pass into working systems.

  • I
    No architecture is sacred
    Model families are hypotheses, not axioms. Any architecture must be revisable when theory or experiment no longer supports it.
  • II
    Explanation before optimization
    Empirical performance alone is not an explanation. Results should be grounded in formal structure, measurable constraints, or identifiable mechanisms.
  • III
    Theory must pass into machinery
    A theoretical result that cannot inform the design of architectures, inference procedures, or computational systems is regarded as incomplete.
  • IV
    Systematic development
    The objective is not a sequence of publications or product cycles. It is a sustained programme that develops theory, models, software, and infrastructure over extended timescales.
What we reject. Treating any fixed architecture as a law of nature. Accepting performance without explanation. Building systems on abstractions that lack formal justification.
003 // Thesis

Scientific motivation

Transformers, diffusion models, and their variants were discovered empirically and then scaled. They work, but they were never derived from physical or mathematical law. The field has powerful systems and no theoretical foundation. ENKAIDU starts from the position that computation and learning are physical processes governed by entropy, energy, and wave dynamics, and that correct architectures can be derived from these laws instead of guessed at and tested. If this derivation succeeds, the resulting systems will not be better versions of current ones. They will be different in kind.

Start from physical law, not architectural convention. Thermodynamics, statistical mechanics, and information theory impose hard constraints on any learning system. These constraints define the structure from which architectures are derived.

Derive, do not guess. Energy landscapes, wave equations, entropy functionals, tensor structures. These are not metaphors. They are the substrates from which model classes and inference procedures are constructed.

Build what the theory produces. Every derivation is expected to yield architectures that work, predictions that can be tested, and systems that can be deployed. If the physics is correct, the engineering follows.

A field that scales what it cannot derive has not yet found its theory.
004 // Research Areas

Research programme

The programme is organised into four areas, each addressing a distinct aspect of the problem: computational substrates, mathematical structure, physical constraints, and higher-order phenomena.

Pillar I

Wave and energy-based computation

Computation governed by wave propagation, energy minimisation, and entropy production, derived from physical law as its own paradigm.

Pillar II

Mathematics of intelligence

Complexity theory, information geometry, topology, and related mathematics applied to characterise the structure and limits of learning and inference.

Pillar III

Physics of computation

Thermodynamic costs, reversibility bounds, and statistical-mechanical analysis of learning, inference, and representation in physical systems.

Pillar IV

Self-reference and higher-order structure

Formal self-reference, reflective computation, and higher-order structures, investigated as structural requirements for systems that reason about their own operation.

005 // Build With Us

Open positions and collaboration

We are looking for researchers in mathematical physics, information theory, and computational science, as well as engineers capable of building the systems this programme requires.

The work spans formal theory, computational experiment, and systems engineering. Candidates should be comfortable operating across more than one of these areas, or deep enough in one to advance it materially.