DEV Community

Kwansub Yun
Kwansub Yun

Posted on

I Built a Physics Verification Engine with Google Gemini. Here's What 700 Tests and 12 Months Taught Me.

Built with Google Gemini: Writing Challenge

This is a submission for the Built with Google Gemini: Writing Challenge

TL;DR

  • I built Flamehaven-TOE, a physics verification pipeline for string-theory background fields.
  • It turns a hypothesis into tensors (G, B, Φ, D) and runs three independent failure channels (BETA / BRST / PDE).
  • Current state: 700 tests passing, 72 skipped, 0 failed (2026-03-04).
  • Gemini’s real value wasn’t “writing code fast” — it was holding equations + Python simultaneously and helping me debug why a test failed.
  • The product outcome is not the model output — it’s the verification infrastructure around it.

1. What I Built with Google Gemini

There are an estimated 10^500 possible universes in string theory's landscape of solutions.

Every one is mathematically self-consistent. Every one describes different physical laws. None can be confirmed by direct observation. We live in one of them, and we don't know which.

This is the problem I have been working on for almost a year.

"The Swampland is not a restriction. It is the universe telling us which mathematics is physics and which is merely mathematics."
— Cumrun Vafa, Harvard University (2005)

Flamehaven-TOE uses Google Gemini as a hypothesis engine, feeding proposed string background configurations into a multi-layer physics pipeline that determines whether a given configuration is physically consistent. It does not claim to find the Theory of Everything. It answers a more tractable question: of the 10^500 candidates, which ones are definitely wrong?

I couldn't find an open-source pipeline that does this at scale. So I built one.

Gemini did not write the tests. Gemini helped write the physics. That distinction matters.


1) 🐈‍⬛The Schrodinger Problem at the Heart of String Theory

Schrodinger's cat is a statement about observation and collapse. Before you open the box, the cat exists in superposition — alive and dead simultaneously. The wave function does not become a particle until someone reaches in to measure it.

The string landscape has exactly this structure.

Every vacuum configuration is in superposition: potentially the right universe, potentially not. Without systematic verification — without reaching into the box — nothing collapses. They all remain equally valid and equally meaningless.

Flamehaven-TOE is the act of reaching into the box.

Each pipeline run is a measurement. Each PASS/FAIL verdict collapses a wave function. The system does not invent physics — it forces candidate configurations to declare themselves.

📷Flamehaven-TOE 0.4.2 dashboard
Flamehaven-TOE dashboard


2) What the Pipeline Does

Topic -> Hypothesis -> BackgroundField -> Beta Residuals -> SIDRCE Gate -> Verdict
           (LLM)       (HypothesisParser)  (A-Module)      (sqrt_jsd)
Enter fullscreen mode Exit fullscreen mode

The engine enforces three independent failure channels — The Laws of the Landscape. These are not software tests; they are physical constraints. If a hypothesis fails here, it is not a bug. It is a mathematical dead end:

Channel Trigger Physics
BETA dilaton gradient, curved spacetime $\beta^\Phi = 4
BRST $D \neq 10$ (superstring) or $D \neq 26$ (bosonic) central charge anomaly
PDE energy growth in coupling evolution dynamical instability

Beyond A-Module, the pipeline includes:

  • B-Module — 4D effective field theory: Kaluza-Klein mass spectrum, Newton constant, and three Swampland conjectures (Distance, Weak Gravity, de Sitter)
  • C-Module — T-duality via Buscher rules and S-duality knowledge graph
  • 5 Physics Agents — quantum invariants, PDE stability, statistical outlier detection, spectral analysis, temporal drift monitoring

"The real challenge... is not proving any single conjecture. It is building a system where multiple independent constraints can be checked simultaneously."
— Tom Banks, Rutgers University


3) Live Results: Four Measurements

📷Pipeline execution yielding a PASS verdict for a mathematically consistent background.
dilaton gradient phi field

📷Pipeline execution catching a beta-function residual and enforcing a FAIL verdict.
WZW S3 exact string solution H-flux


These are not mocked results. Actual pipeline output, executed 2026-03-04 @ commit 4848120. Four wave functions forced to declare themselves:

  Topic                              beta_Phi    beta    brst  omega   gate
  -------------------------------------------------------------------------
  WZW S3 exact solution              0.00e+00    pass    True  1.0000  PASS
  Dilaton gradient phi field         4.69e-01    fail    True  0.0000  FAIL
  Schwarzschild near-horizon         9.90e-03    border  True  0.8351  FAIL
  D=4 unification                    0.00e+00    fail   False  1.0000  FAIL
Enter fullscreen mode Exit fullscreen mode

Four topics. Four distinct failure signatures.

The WZW S^3 model passes because Ricci curvature and H-flux torsion cancel to machine precision — the canonical non-trivial string background from Polchinski §15.1. This is not a trivial PASS. The cancellation $R_{ab} = \frac{1}{4}H_{acd}H_b^{\;cd}$ requires the full Christoffel $\rightarrow$ Riemann $\rightarrow$ Ricci tensor chain, plus the H-flux contraction, to be algebraically correct. It is the hardest test in the suite, and the most meaningful one.

Why RAG Fails the WZW Model:
In the WZW $S^3$ model, the cancellation of curvature and torsion is a global property of the manifold. A RAG system might retrieve the definition of the Ricci tensor in one chunk and the H-flux in another, but it cannot "see" the algebraic requirement that $R_{ab} = \frac{1}{4}H_{acd}H_b^{\;cd}$ across the entire codebase. Gemini 3.1 Pro’s long-context reasoning allows it to hold the entire geometric identity in a single "Epistemic Map," verifying the physics instead of just pattern-matching the syntax.

The Schwarzschild case is borderline — caught by the beta gate (omega=0.8351) even though SIDRCE passes. The measurement apparatus is precise enough to see the boundary.

The D=4 case fails via BRST — wrong spacetime dimension for superstring theory, caught independently of beta functions entirely. The universe doesn't care about your metric if the dimension is wrong.


4) Scientific Validation

"A verification engine for string backgrounds... is itself a form of theoretical discovery. The act of elimination is the act of measurement."
— Erik Verlinde, University of Amsterdam

Validated against published exact solutions:

Background Reference beta residual Gate
10D Minkowski (G=eta, Phi=const) Polchinski §3.7 0.0 PASS
Linear Dilaton (Phi = V·x) Polchinski §3.4 4V^2 (non-zero) FAIL
WZW S^3 (R_ab = H^2_ab cancellation) Polchinski §15.1 0.0 PASS

Layer 2 cross-validation: Christoffel->Riemann->Ricci numerical chain independently verified against sympy symbolic computation on S^2 geometry. Max deviation < 1e-3.


5) What the Slop Audit Actually Found

I have the full audit from v3.1.0 — 77 files, measured in production:

v3.1.0 Slop Audit — Measured Values
  files_audited:          77
  status_distribution:    clean=66  dependency_noise=10  suspicious=1
  ldr_grades:             77/77 (min=0.909, max=1.000, mean=0.996)
  files_below_ldr_085:    0
  actual_hallucinations:  2 out of 77 files  (2.6%)
  avg_inflation:          0.019
Enter fullscreen mode Exit fullscreen mode

11 deficit files = 10 style-only (__future__/typing unused imports — Python boilerplate, not AI hallucinations, zero fake imports) + 1 suspicious (pennylane_engine.py: torch hallucination + 2 empty ImportError handlers that silently swallow physics failures). The second flagged file, a numpy import in a patch file, was a DDC false positive.

The highest inflation score: singularity_bridge.py at 0.646, flagged because "equation" appears 21 times. Reviewed manually: confirmed clean. In a physics codebase, "equation" is not jargon — it is the work. Physics-domain slop detection requires domain-aware thresholds, and that gap is a limitation of the current tooling.

The physics core files — the ones that matter most:

beta_residual.py       ldr=0.995  inflation=0.000
tensors.py             ldr=1.000  inflation=0.000
duality_kg.py          ldr=1.000  inflation=0.050
cy_bridge.py           ldr=1.000  inflation=0.000
eft_generator.py       ldr=0.996  inflation=0.000
hypothesis_parser.py   ldr=0.990  inflation=0.000
Enter fullscreen mode Exit fullscreen mode

All six: inflation zero or near-zero. The files that run the physics are the cleanest in the codebase. The gate works where it matters.


2. Where Gemini Built the Physics

Gemini's role was not code generation in the "write me a function" sense. It was formula-to-code translation with simultaneous algebraic verification — holding a mathematical formula and a Python target in the same working context, in consistent index notation, and producing code that is algebraically faithful to the physics.

The T-duality transformation via Buscher rules was the clearest example. I gave Gemini the 1987 Buscher rules for how metric, B-field, and dilaton transform under T-duality along a Killing direction. Gemini held the full set of transformation equations and produced Python code that correctly contracts indices via einsum:

# C-Module: T-duality via Buscher rules
# The negative sign in the dilaton shift took 3 sessions to stabilize.

def apply_t_duality(bg: BackgroundField, direction: int) -> BackgroundField:
    a = direction
    G, B, phi = bg.G, bg.B, bg.phi
    G_aa = G[a, a]

    # Buscher rules (Polchinski convention):
    # G'_aa = 1/G_aa
    # G'_ai = B_ai / G_aa
    # G'_ij = G_ij - (G_ia*G_ja - B_ia*B_ja) / G_aa
    # Phi'  = Phi - 0.5 * ln(|G_aa|)     <- THIS SIGN took 3 sessions
    ...
Enter fullscreen mode Exit fullscreen mode

That negative sign — Phi - 0.5 * ln(|G_aa|) — appeared as positive in two out of three sessions because Gemini picked different conventions from different papers in its training data. This specific bug is why convention contracts exist.


1) Version Trajectory: What Changed — And Why

The project's evolution mirrors Gemini's own model evolution. Three model eras, three qualitative shifts.

Era 1 — Gemini 2.5 Pro (March 2025 – November 2025): Foundation

v1.0 → v2.5    Gemini 2.5 Pro (released 2025-03-25)
  Core pipeline: Harvester → Reasoner → SR9/DI2/Gate → Store
  Text-mode only (GOLD/SILVER/REJECTED scoring)
  First convention contracts, first DDC gate
  Wall: 1M token context was enough for pipeline logic,
        but not enough for simultaneous physics + code + verification
Enter fullscreen mode Exit fullscreen mode

This era built the skeleton. Gemini 2.5 Pro — the first "thinking model" with extended reasoning — could generate individual modules, but could not hold the full physics specification (beta functions + Buscher rules + Swampland conjectures) in a single session and produce consistent code. Each module worked. They did not talk to each other correctly. The wall was not context length. It was reasoning depth: the model could not chain physics constraints across modules without losing consistency.

Era 2 — Gemini 3 Pro (November 2025 – February 2026): Physics Depth

v2.5 → v3.0    Gemini 3 Pro (released 2025-11-18)
  Physics-mode pipeline: HypothesisParser → BetaResidualVerifier → SIDRCE
  A-Module (beta functions), B-Module (Swampland DC), C-Module (T-duality)
  3 gate cases: Minkowski PASS, linear dilaton FAIL, D=4 BRST FAIL
  Wall: could hold the spec, but lost index structure mid-derivation
        (WZW cancellation required multi-session manual correction)
Enter fullscreen mode Exit fullscreen mode

Gemini 3 Pro — marketed as "most intelligent model" with native tool use and deeper multi-step reasoning — broke through the first wall. It could hold the full specification in context and reason across modules. The physics pipeline came alive: hypotheses went in, verdicts came out. But the second wall appeared: multi-step tensor derivations. The WZW S^3 cancellation, where R_{ab} = (1/4)H_{acd}H_b^{cd}, required getting the Christoffel->Riemann->Ricci chain plus the H-flux contraction correct simultaneously. At 3 Pro, Gemini would get the Christoffel symbols right but lose the H-flux index structure mid-derivation. The reasoning was deep enough to start the chain but not to finish it without human correction.

Era 3 — Gemini 3.1 Pro (February 2026 – present): Gate Differentiation

v3.1.0  (2026-02-22)  <-  Gemini 3.1 Pro (released 2026-02-19, adopted within 3 days)
  77 files | 3 gate cases | 2 hallucinations (2.6%) | ~50 tests
  Swampland: Distance Conjecture only

v3.4.0  (2026-02-23)  <-  Gemini 3.1 Pro
  328 tests | +Swampland WGC + de Sitter conjectures
  pyCICY: 3 geometry validations | T/S-duality self-consistency gates
  Breakthrough: reasoning depth allowed B-Module and C-Module in single session

v4.0.2  (2026-03-04)  <-  Gemini 3.1 Pro
  700 tests (+113% from v3.4.0) | +WZW S^3 exact solution (non-trivial PASS)
  +SIDRCE omega scoring | +5 physics agents + governance
  +PDE stability channel | Web dashboard (Next.js + FastAPI)
  Breakthrough: WZW cancellation (R_ab = H^2_ab) implemented in single pass
Enter fullscreen mode Exit fullscreen mode

Gemini 3.1 Pro — with its enhanced reasoning and agentic capabilities — broke the second wall. The full tensor chain came out algebraically correct in a single pass. This was the "smoking gun" for reasoning coherence. The model successfully maintained the index structure across the Ricci contraction and the H-flux, producing code like this without human intervention:

# WZW S^3 Exact Cancellation (Generated by 3.1 Pro)
# R_ab - (1/4) H_acd H_b^cd = 0
H2_ab = np.einsum('acd,bcd->ab', H_flux, H_flux) * G_inv_tensor
beta_G = R_ab - 0.25 * H2_ab
Enter fullscreen mode Exit fullscreen mode

Each era shifted the bottleneck: Can Gemini do this at all?Can it hold enough context?Can it maintain algebraic coherence across a full derivation? The last question is the one 3.1 Pro answered.

One number tells the parser improvement story: the linear dilaton $\beta^\Phi$ went from 2.57e-04 (v3.1.0) to 4.69e-01 (v4.0.2). Not a regression. The HypothesisParser learned to extract a larger gradient amplitude from the same text — 4 * (0.342)^2 = 0.468. Better parser produces a stronger, more physically accurate failure signal.


2) Current State

pytest

700 tests passing, 72 skipped, 0 failed. Version 4.0.2:

src/toe/
  engine/        Pipeline orchestrator + pydantic-settings config
  physics/       A-Module (BetaResidualVerifier, HypothesisParser)
                 B-Module (EFT4DGenerator: KK spectrum, Swampland DC/WGC/dS)
                 C-Module (DualityKGIntegrator: T-duality, S-duality)
                 cy_bridge (CYTools/pyCICY/cymyc/JAX Yukawa adapters)
  sidrce/        sqrt_jsd gate + omega scoring
  governance/    5 physics agents + orchestrator
  math/          SR9, DI2, Gate eval + Rust FFI bridge
  drift/         Temporal drift monitoring
Enter fullscreen mode Exit fullscreen mode

Web dashboard (Next.js + FastAPI) with full feature parity to the Rich TUI. Interactive QNE scan, drift visualization, export MD/CSV, AI-enhanced reporting via Gemini API.


3. What I Learned

1) The Difference Between "Code That Compiles" and "Code That Is Physically Correct"

In conventional software, a passing test suite means the code works. In physics code, it means the code is internally consistent — much weaker. Code can be self-consistent and physically wrong. The wave function can look collapsed when it hasn't been measured at all.

Three verification layers emerged, each from a specific failure:

DDC (Dependency Density Check) — Catches hallucinated imports. The trajectory is honest: at v3.1.0 (77 files), the actual hallucination rate was already 2.6%. But that number hides the DDC's own limitations — 10 false positives on __future__/typing imports, and a physics-domain false alarm on singularity_bridge.py where "equation" appearing 21 times triggered the inflation score. A general slop detector has domain limitations. But the physics core files — the six that run actual tensor computations — all scored inflation=0.000. The gate works where it matters.

Convention contracts — Machine-enforced parser trigger tables. Every hypothesis keyword maps to a specific BackgroundField parameter and an expected gate result. Tested in test_parser_contracts.py. Eliminates the bug class where "the AI gave different index conventions in different sessions."

Two-layer scientific validation — Exact solutions from textbooks (Layer 1) + independent sympy cross-validation (Layer 2). Born from discovering Gemini could produce code that passed unit tests but used the wrong sign for the Buscher dilaton shift.

2) The 20% Tax

The most expensive operational cost: context reconstruction consumed ~20% of every new session. Not because Gemini forgets — because session boundaries are hard resets.

The T-duality sign appeared in three forms across the codebase. Each form was valid in some paper. Each came from a different Gemini session. Three different measurements of the same cat — each internally consistent, each looking at a different box.

The fix: a CONTEXT.md with established conventions, validated approximations, and open stubs. Every session begins with that context. The tax dropped to under 5%.

3) Confidence Calibration Is Flat

I expected a fast coder that occasionally gets physics wrong. What I found: Gemini's physics reasoning is often correct, but every claim — exact result, standard approximation, plausible guess — is presented at the same confidence level. In physics code, this is the root cause of half the bugs.

The lesson: the value of an AI physics partner is the verification infrastructure you build around it. DDC, convention contracts, two-layer validation, 5-agent governance — these are the product. Gemini is the engine. The pipeline is the vehicle. The measurement apparatus matters more than the particle source.


4. Google Gemini Feedback

Evidence baseline:

  • 700 tests (700 passed, 72 skipped, 0 failed — 2026-03-04)
  • v3.1.0 slop audit: 77 files, 2.6% hallucination rate, physics core inflation=0.000
  • 12 months continuous development, Gemini 2.5 Pro through 3.1 Pro

Be candid. So here it is — candid.

1) What Worked

Long-context reasoning over short-context retrieval. In theoretical physics, local logic is useless without global consistency. Traditional RAG breaks a manifold's topology into disconnected chunks — a Christoffel symbol in one chunk, the Bianchi identity that constrains it in another, the Swampland conjecture that eliminates the solution in a third. When Gemini held the full Flamehaven-TOE specification in a single context, it didn't just search for relevant formulas — it reasoned across the entire mathematical landscape simultaneously. This is the single capability that made the project possible. RAG would have produced locally coherent, globally inconsistent physics.

Simultaneous code and formula coherence. Holding a formula and a Python target simultaneously — correct index notation, correct einsum contractions — and producing code that is algebraically faithful. Unreliable at month 1; the foundation of the entire pipeline now.

Reasoning about failure. When I showed Gemini a failing test alongside the formula it violated, it could trace the physical meaning of the index error back to the formula — not pattern-match to a fix. This is what makes it valuable as a hypothesis engine.

The model evolution trajectory matters. The upgrade path from Gemini 2.5 Pro (March 2025) through 3 Pro (November 2025) to 3.1 Pro (February 2026) was not a smooth line — it was a series of specific walls coming down. With 3 Pro, implementing the WZW S^3 cancellation required multiple sessions of manual index correction. With 3.1 Pro, the full tensor chain — Christoffel symbols, Riemann tensor, Ricci contraction, plus the H-flux term $(1/4)H_{acd}H_b^{\;cd}$ — came out algebraically correct in a single pass. That is measurable progress. It shifted the bottleneck from "can the model produce correct tensor algebra" to "how do I verify what it produced" — a qualitatively better problem.


2) What Still Creates Friction — Three Distinct Failure Modes

Physics-domain hallucination is not one problem. It is three problems with three distinct causes. The distinction matters because the fixes are different.

Failure Mode 1: Library Hallucination = API Grounding Failure

Gemini describes pyCICY, cymetric, or cymyc functions that don't exist — confidently, without uncertainty markers. The cause: no runtime grounding in what package APIs actually expose. v3.1.0 measured: 2 hallucinated dependencies out of 77 files (2.6%). Both caught by DDC before reaching the physics pipeline.

# v3.1.0, hallucinated:
import torch  # in pennylane_engine.py — PyTorch not in this environment
# DDC caught it. Physics pipeline never saw it.
Enter fullscreen mode Exit fullscreen mode

Failure Mode 2: Approximation Without Marking = Epistemic Labeling Failure

LLMs are born dreamers. In a physics engine, a "dream" is a violation of conservation laws. When Gemini approximates $\nabla^2\Phi \approx 0$ for a linear dilaton, the result is presented as exact — no validity conditions, no breaks-when, no confidence qualifier. The output format has no slot for epistemic metadata.

My solution: I stopped asking Gemini for answers and started asking for verifications. By feeding the entire Flamehaven-TOE blueprint as context, I forced the model to reason against my own metric_library.py and hypothesis_parser.py before making a claim. This turned hallucination into structured peer review. But the output format still needs a native mechanism:

{
  "claim": "nabla^2 Phi = 0 for linear dilaton",
  "type": "approximation",
  "validity_condition": "flat background, linear Phi(x)",
  "breaks_when": "curved background, non-linear dilaton",
  "confidence": "high_within_stated_conditions"
}
Enter fullscreen mode Exit fullscreen mode

Failure Mode 3: Convention Drift = Project-Level State Failure

Gemini switches Riemann tensor conventions mid-derivation — syntactically valid, physically inconsistent. ~50% of bugs caught in review. The cause: no persistent project state carrying "in this project, we use these conventions." Fix is project memory, not more training.


3) What Is Genuinely Needed: Four Unlocks

Unlock 1: Colab + Managed Physics Environment

The Flamehaven-TOE Phase 4/5 blockers — CYTools, cymetric/cyjax, cymyc — all require either Docker containers or Linux/WSL2 environments. None run in a standard Python environment. All are currently stubs.

CYTools targets Colab as a primary platform and provides a prebuilt container image. cymetric and cyjax are TensorFlow/JAX-based and run on Colab GPU/TPU natively. The integration path exists. What is missing is a managed execution environment where Gemini can trigger a Colab runtime, run the physics computation, and return structured results back into the pipeline.

# This is what Phase 4 looks like if the managed environment exists:

# CYTools via Colab managed runtime
from cytools import Polytope
p   = Polytope([[1,0,0,0],[-1,1,0,0],[0,-1,1,0],[0,0,-1,1],[0,0,0,-1]])
cy  = p.get_cy()
h11, h21 = cy.h11, cy.h21   # real Hodge numbers from FRST triangulation
                              # not pyCICY approximations

# cymetric via Colab GPU
from cymetric.models.tfmodels import PhiModel
model = PhiModel(cy)
model.train(epochs=100)      # Ricci-flat metric — replaces unit Kahler stub
Enter fullscreen mode Exit fullscreen mode

If Gemini could execute this in a managed Colab environment, Phase 4 becomes runnable — not theoretical. The topology_gap metric stops being a stub and starts being a measurement.

Unlock 2: Domain-Locked Corpus Mode for Physics

The theoretical physics ecosystem is heavily fragmented by competing index conventions and notation standards. The solution is not fine-tuning, but the creation of a domain-locked "Digital Library of Alexandria" for physics execution — a mode where Gemini's code generation is strictly grounded against a curated, canonical corpus:

Domain-Locked Corpus / Retrieval Context:
  Canonical textbooks:
    Polchinski, "String Theory" Vol. 1-2
    Green, Schwarz, Witten, "Superstring Theory" Vol. 1-2
    Becker, Becker, Schwarz, "String Theory and M-Theory"

  Swampland review:   van Beest et al. arXiv:2109.06925
  T-duality:          Buscher (1987) Phys.Lett.B201 (original)
  TASI flux lectures: Denef, Douglas, Kachru

  Package documentation + source:
    pyCICY v0.5.2
    CYTools latest
    cymyc arXiv:2410.19728
    cymetric arXiv:2111.01436
Enter fullscreen mode Exit fullscreen mode

With this retrieval context active, the index convention question — "what is the standard sign of the Buscher dilaton shift in Polchinski's notation?" — has a specific, retrievable answer. Convention drift drops to near zero because the answer comes from a single grounded source, not from aggregating across incompatible paper conventions.

This is distinct from fine-tuning. It is a retrieval-augmented execution mode. The model's weights do not change; the lookup corpus does.

Unlock 3: Structured Uncertainty Output

Described above under Failure Mode 2. The output structure addition — claim type, validity condition, breaks-when, reference — is the highest-leverage single change for physics code generation. It converts approximations from invisible risks to documented design decisions.

Unlock 4: Persistent Project State Across Sessions

Convention drift across sessions is a measurable cost — approximately 20% of every new session was spent reconstructing established context. The same T-duality sign convention appeared in three forms across the codebase before being unified.

What is needed is not conversation memory. It is structured physics project state:

{
  "project": "flamehaven-toe",
  "established_conventions": {
    "riemann_tensor": "R^rho_{sigma mu nu} — upper first index",
    "buscher_dilaton": "Phi' = Phi - 0.5 * ln(G_aa) — negative shift",
    "metric_signature": "(-,+,+,...,+) — mostly plus"
  },
  "validated_approximations": [
    {
      "claim": "nabla^2 Phi = 0",
      "valid_when": "flat background, linear dilaton",
      "validated_in": "session_2026_02_15"
    }
  ],
  "open_stubs": ["yukawa_couplings", "cy_metric", "ads_cft_duality"]
}
Enter fullscreen mode Exit fullscreen mode

Loaded at session start, updated at session end. Convention drift eliminated. Approximation history preserved. Open stubs visible. Multi-session physics research becomes qualitatively different.


4) The Vision

Researcher: "Scan the Swampland WGC bound
             as g_s varies from 0.01 to 0.5 across the CICY dataset."

Gemini + Managed Colab + Domain Corpus + Persistent State:
  1. Loads project conventions (no drift)
  2. Spins up CYTools in managed environment
  3. Generates 500 BackgroundField configurations
  4. Runs each through A/B/C pipeline
  5. Returns: PASS/FAIL heatmap, topology_gap distribution, WGC boundary
  6. Flags: "3 geometries survive all gates at g_s in [0.08, 0.12]"
  7. Epistemic labels on every intermediate claim
Enter fullscreen mode Exit fullscreen mode

Every component either exists or is one unlock away. The pipeline is built and passing 700 tests. The gap is exactly these four capabilities.


5. What I'm Building Next

The pipeline is gated and tested. The next step is to make the physics stubs runnable end-to-end:

  • Managed execution for geometry toolchains (CYTools, cymetric, cyjax)
  • Domain-grounded retrieval for convention stability across sessions
  • Artifact-first outputs — audit trails, reproducible results, tamper-resistant verification logs

If you're building anything where correctness is not negotiable, I'd love to compare notes.

The cat is still in the box. The measurement apparatus is more precise than it was a year ago.

Every run collapses one more wave function. The landscape is vast. We are beginning to map it.

Twelve months of trajectory — and the concrete progress from Gemini 2.5 Pro through 3 Pro to 3.1 Pro — says the tools will grow to meet the problem.


From someone who has used Gemini as a physics research tool for twelve months, watched it grow from 2.5 Pro through 3 Pro to 3.1 Pro while the project grew from 50 to 700 tests, measured the hallucination rate at 2.6% and the physics core inflation at zero, and wants it to become what it is clearly capable of becoming.

Top comments (0)