Recursive Golden‑Ratio Decision Systems

Comprehensive benchmarks and simulations across five optimization dimensions using φ‑based frequency‑state models.

φ≈1.618 hi=1/φ≈0.618 lo=1/φ²≈0.382

Introduction

The golden ratio φ (approximately 1.618) appears in contexts ranging from quasicrystals and phyllotaxis to algorithmic search and control theory. Its continued fraction representation [1;1,1,1,…] makes it the "most irrational" number, which prevents simple rational resonances. This property underpins many stability and optimisation results. Here we synthesise the mathematical, economic, engineering and biological appearances of φ, and build a unified frequency‑state decision model that generalises golden‑ratio search. We then compare this Golden Gradient approach with traditional optimisers across five comprehensive benchmark dimensions.

Mathematical Foundations

Hurwitz's theorem shows that φ is the worst approximable number, giving it a special place in Diophantine approximation. Zeckendorf's theorem states that every integer can be uniquely represented as a sum of non‑consecutive Fibonacci numbers. These two results link φ to optimal search: golden‑section search divides an interval in the ratio 0.618 to 0.382 at each step and minimises the worst‑case number of evaluations for unimodal functions.

In control theory, Benavoli et al. proved that the steady‑state Kalman gain of a scalar random‑walk system with equal noise variances converges exactly to hi=0.618, with the corresponding error covariance approaching φ. This implies that optimal linear fusion weights between prior estimate and new measurement follow the golden ratio. In linear– quadratic control and inventory smoothing, φ arises as the optimal adjustment factor.

Theoretical Snapshot

We anchor the golden‑ratio dynamics with \( \varphi = \frac{1 + \sqrt{5}}{2} \), and its reciprocal \( \varphi^{-1} \approx 0.618 \). Each state \( s \) maintains a Laplace‑smoothed success rate \( r_s = \frac{a_s + 1}{a_s + b_s + 2} \), which determines whether we follow or flip the nominal golden‑section move.

The canonical step update for a bounded one‑dimensional slice uses a signed direction \( \sigma_t \in \{-1, +1\} \) and adaptive scale \( \eta_t \):

\[ x_{t+1} = x_t + \sigma_t \, \varphi^{-1} (b-a) \, \eta_t \]

To detect regime changes, we compare recent and baseline state frequencies with a Wasserstein distance:

\[ p_{\mathrm{drift}} = W_1(\text{recent}, \text{baseline}) \]

Applications Across Domains

Finance. Fibonacci retracements are widely used in technical analysis to draw support and resistance levels at 23.6%, 38.2%, 61.8%, etc., though evidence for predictive power is mixed. More compelling are φ‑based allocation strategies: portfolios with asset weights in 1:φ ratios show robust performance over decades, and corporate capital structures near φ proportion outperform random allocations.

Machine learning. Modern hyperparameter and neural architecture search methods implicitly partition spaces into "good" and "bad" regions. Golden‑ratio proximal algorithms achieve large step sizes and provable convergence. Golden Gradient extends this by treating the entire search process as a Markov decision problem with φ‑based priors.

Signal processing. φ appears in optimal Kalman gains and in the organisation of neuronal oscillations. Frequency ratios separated by φ minimise cross‑frequency interference. Fibonacci lattice sampling yields near‑uniform point distributions on the sphere, reducing error versus latitude–longitude grids.

Game theory and biology. Ultimatum game experiments show that offers around 38.2% maximise acceptance probability—matching the golden split. Phyllotaxis patterns in sunflowers, pine cones and cacti place successive leaves at 137.5°, the "golden angle", optimising sunlight and packing efficiency. Self‑organising dynamical systems recreate these patterns from simple rules.

Computer science and operations research. Fibonacci heaps achieve O(1) decrease‑key operations with tree heights bounded by logφ(n). Multiplicative hashing uses φ to distribute keys uniformly. In supply chains, the golden smoothing rule suggests adjusting inventory by 61.8% of the discrepancy each period to minimise variance.

The Frequency‑State Model

Golden Gradient generalises golden‑section search by framing optimisation as navigation on a frequency‑state graph. At each iteration, a decision is made to cut the search region in a φ ratio (hi≈0.618 vs lo≈0.382). The algorithm records a history of long (L) or short (S) decisions, forming a discrete state sequence. Each state s has an associated success probability rs, estimated from past outcomes. When facing a new decision, if rs≥0.5 the algorithm follows the nominal golden‑section recommendation; if rs<0.5 it flips direction. Periodic verification steps re‑evaluate previously discarded options to detect drift. A Wasserstein distance between recent and historical state distributions flags regime changes and triggers extra verification.

φ = \frac{1+\sqrt{5}}{2} ≈ 1.618 hi = \frac{1}{φ} ≈ 0.618 lo = \frac{1}{φ^2} ≈ 0.382 r_s = \frac{a_s + 1}{a_s + b_s + 2}\qquad \text{(Laplace‑smoothed success probability)} p\_\mathrm{drift} = W\_1\bigl(\text{recent state frequencies},\ \text{baseline frequencies}\bigr)

This frequency‑state approach retains the interval‑reduction optimality of golden‑section search while learning from context. It naturally incorporates additional optimisers (CMA‑ES, Nelder–Mead, Bayesian optimisation, Particle Swarm, Differential Evolution) by allowing them to operate as modules that propose candidate points. At each step(), external proposal generators are called and their candidate positions are immediately clamped to the active bounds. These external candidates then pass through the same frequency‑state key construction, state‑probability lookup, and Aharonov‑Bohm phase adjustment as native moves, with a source label attached for debugging.

Theoretical Perspectives on GoldenGradient through Measure Theory and Duality

Introduction

GoldenGradient can be reframed as a probabilistic decision process that evolves over a structured state space. Measure theory offers a clean vocabulary for describing how state frequencies, update rules, and drift detection behave under uncertainty.

Algorithm Overview

The algorithm maintains a history of frequency‑state outcomes and updates its policy based on observed success rates. Each iteration chooses a golden‑ratio cut, optionally flips direction, and performs verification checks when drift is detected.

Measure‑Theoretic Framing

The frequency‑state space can be treated as a measurable space (𝒮, 𝔽), where 𝒮 collects all discrete states and 𝔽 is the σ‑algebra over those states. Each iteration induces a probability measure μ over 𝒮, capturing how often the algorithm visits each state.

Probability & Integration

Success probabilities are updated with a Beta prior and a Laplace‑smoothed posterior mean, so integration over μ becomes a simple weighted average of state outcomes.

Wasserstein Distance

Drift detection compares recent and historical measures using the 1‑Wasserstein distance, highlighting transport cost between two frequency distributions. Large distances trigger verification actions and additional sampling.

Duality & Representation

Riesz representation links linear functionals on continuous functions over 𝒮 to measures, providing a principled way to interpret score functions as integrals. This viewpoint aligns with MDP duality, where value functions and occupation measures describe the same policy in primal and dual forms.

Computational Implications

The measure‑theoretic framing clarifies why small updates remain stable and why Wasserstein drift detection is robust to noisy transitions. It also motivates efficient summaries of μ, such as bucketed histograms and sparse state visitation.

Metaheuristic Positioning

Golden Gradient sits between deterministic interval methods and population‑based metaheuristics. It preserves golden‑section optimality for unimodal slices while delegating exploration to external proposal modules (CMA‑ES, Nelder–Mead, Bayesian optimisation, Particle Swarm, Differential Evolution), then adjudicates their candidates using the learned frequency‑state policy.

x\_\mathrm{cand} \sim \mathcal{M}\_\mathrm{proposal}(\Omega) a(x\_\mathrm{cand}) = \mathbb{I}[r\_s \ge 0.5]\ \lor\ \mathbb{I}[\text{verify}(s)] x\_{t+1} = \arg\min\_{x \in \{x\_\mathrm{cand}, x\_\mathrm{verify}\}} f(x)

This positioning allows Golden Gradient to remain sample‑efficient on smooth objectives while still leveraging broader exploration when the proposal modules detect non‑convex structure.

Multi‑Objective Variants

Multi‑objective objectives are handled through either scalarization or Pareto‑based selection. Scalarization compresses multiple criteria into a single target, while the Pareto‑based path retains a Pareto set of non‑dominated proposals for downstream MCDA selection.

f\_\lambda(x) = \sum\_{k=1}^{K} \lambda\_k f\_k(x), \qquad \sum\_k \lambda\_k = 1 f\_\infty(x) = \max\_k \lambda\_k |f\_k(x) - z^\*_{k}| \mathcal{P} = \{x : \nexists y \text{ such that } f(y) \preceq f(x)\}

The Pareto set hook stores non‑dominated proposals emitted by the proposal modules, enabling later MCDA scoring without discarding trade‑off structure.

MCDA Selection Strategies

When multiple objectives are active, the system applies MCDA scoring to choose among candidates in the Pareto set. The MCDA hook is deliberately modular so new scoring rules can be added without altering proposal generation.

u(x) = \sum\_{k=1}^{K} w\_k \cdot \hat{f}\_k(x) s(x) = \alpha \cdot u(x) + (1-\alpha) \cdot d(x, z^\*) x^\* = \arg\max\_{x \in \mathcal{P}} s(x)

MCDA scoring consumes the Pareto set hook and produces the final accepted candidate, which then feeds back into the frequency‑state update for the next iteration.

Multi‑Dimensional Applications

Golden Gradient can be extended to multiple variables by treating each dimension as its own search axis. At each iteration the algorithm cuts along each dimension in the golden ratio and records whether the move was on the high side (L) or low side (S). The collection of decisions across all dimensions forms a discrete joint state vector represented by a tuple (s1, s2, …, sd). Each state vector has an associated success probability r(s1,…,sd) estimated from past outcomes, and future decisions can either follow or flip the golden‑ratio recommendation based on these probabilities.

Dimensional Reduction Formulas

The algorithm's complexity and power emerge from the recursive structure across dimensions. We express the core formulas for each dimensional layer, showing how n-dimensional search decomposes into lower-dimensional cross-sections. Let φ = (1+√5)/2 denote the golden ratio.

Definition 1 (State Space). For an n-dimensional search space Ω ⊂ ℝⁿ, the discrete state space is: 𝒮ₙ = 𝒯ₙ × 𝒲 × 𝒞ₙ where: 𝒯ₙ = {0, 1, 2}ⁿ (frequency discretization, 3 buckets per dimension) 𝒲 = {0, 1, 2, 3} (Wasserstein distance buckets) 𝒞ₙ = {0, 1, 2, 3}^C(n,2) (cross-section distance buckets) |𝒮ₙ| = 3ⁿ · 4 · 4^(n(n-1)/2) Definition 2 (Frequency State). For position x ∈ Ω with bounds [aᵢ, bᵢ], the frequency state is: fᵢ(x) = (xᵢ - aᵢ)/(bᵢ - aᵢ) ∈ [0,1] for i = 1, ..., n with discretization τ: [0,1] → {0,1,2}: ⎧ 0, if f < φ⁻² ≈ 0.382 τ(f) = ⎨ 1, if φ⁻² ≤ f < φ⁻¹ ≈ 0.618 ⎩ 2, if f ≥ φ⁻¹ Definition 3 (Cross-Section Distance). For dimension pair (i,j) with i < j: dᵢⱼ(x) = |fᵢ(x) - fⱼ(x)| The set of all cross-sections forms the complete graph K_n on n vertices: ℰₙ = {(i,j) : 1 ≤ i < j ≤ n}, |ℰₙ| = C(n,2) = n(n-1)/2 Theorem 1 (Dimensional Decomposition). The n-dimensional state space decomposes recursively: 𝒮ₙ = 𝒮ₙ₋₁ ⊗ 𝒯₁ ⊗ 𝒞ₙ₋₁,ₙ where 𝒞ₙ₋₁,ₙ = {0,1,2,3}^(n-1) represents cross-sections involving dimension n. Proof. The cross-sections partition as ℰₙ = ℰₙ₋₁ ∪ {(i,n) : i < n}, giving |ℰₙ₋₁| + (n-1) = |ℰₙ|. □

Phase Dynamics (Aharonov-Bohm Formulation)

Definition 4 (Vector Potential). Define the learned potential field V: Ω → ℝ as: V(x) = (1/Nₓ) Σₖ v(xₖ) · 𝟙[xₖ ∈ B(x,ε)] where B(x,ε) is an ε-ball and Nₓ counts observations in that ball. Definition 5 (Topological Phase). Along a path γ: [0,T] → Ω, the accumulated phase is: T Φ[γ] = ∫ A(γ(t)) · γ̇(t) dt 0 where the connection 1-form A is defined componentwise: Aᵢ = (2π/φ²) · (∂V/∂xᵢ)/(|V| + 1) The golden angle 2π/φ² ≈ 137.5° ensures maximal irrationality, preventing resonance. Definition 6 (Winding Number). For a closed path γ encircling region R: W[γ] = (1/2π) ∮ A · dl = (1/2π) ∫∫ (∇ × A) · dS γ R In discrete form, when ‖x(t) - x(s)‖ < ε for t - s > δ: W ← W + sign(v(x(s)) - v(x(t))) Proposition 1 (Phase-Adjusted Probability). The selection probability transforms as: p'(s) = p(s) + λ₁ · (V(xₛ) - V(x_current))/(|V(x_current)| + 1) + λ₂ · sign(W) with hyperparameters λ₁ = 0.1 (potential weight) and λ₂ = 0.05 (winding weight).

Explicit Formulas for Dimensions 1–5

n = 1 (Base Case): |ℰ₁| = 0, |𝒮₁| = 3 · 4 · 1 = 12 xₖ₊₁ = xₖ + σₖ · sign(ξₖ) · φ⁻¹ · (b - a) · ηₖ where σₖ ∈ {-1,+1} is direction, ξₖ ~ Bernoulli(p(sₖ)), and ηₖ is adaptive step scale. n = 2: |ℰ₂| = 1, |𝒮₂| = 9 · 4 · 4 = 144 Φ₂ = Σₖ [(Δvₖ)/(|vₖ|+1) + (Δx₁,ₖ + Δx₂,ₖ) · (2π/φ²)] n = 3: |ℰ₃| = 3, |𝒮₃| = 27 · 4 · 64 = 6,912 ℰ₃ = {(1,2), (1,3), (2,3)} Φ₃ = Φ₂ + ∮ A₁₃ · dl + ∮ A₂₃ · dl n = 4: |ℰ₄| = 6, |𝒮₄| = 81 · 4 · 4,096 = 1,327,104 ℰ₄ = {(1,2), (1,3), (1,4), (2,3), (2,4), (3,4)} W₄ ∈ ℤ, Φ₄ = Φ₃ + Σ ∮ Aᵢⱼ · dl (i,4)∈ℰ₄ n = 5: |ℰ₅| = 10, |𝒮₅| = 243 · 4 · 4¹⁰ ≈ 1.02 × 10⁹ A = (A₁₂, A₁₃, A₁₄, A₁₅, A₂₃, A₂₄, A₂₅, A₃₄, A₃₅, A₄₅) ∈ ℝ¹⁰ Φ₅ = ∫ A · dx, where dx ∈ T*Ω is the cotangent bundle γ

To understand decision‑making in truly multidimensional settings, consider a 4D hypercube where each axis represents a distinct marginal state distribution. Each marginal applies Golden Section Search (GSS)—the frequency distribution itself is a golden‑ratio distribution where the center holds the most probability mass. Every vertex in the hypercube corresponds to a unique joint state vector across four dimensions—for instance, (L,S,L,S) indicates upper-partition decisions on dimensions 1 and 3, and lower-partition decisions on dimensions 2 and 4. Each vertex carries an empirical probability mass derived from historical observations.

In higher‑dimensional spaces, dimensional cross‑sections reveal a crucial phenomenon: rather than a single upper-tail regime and a single lower-tail regime, the joint distribution exhibits multiple upper-tail regimes and multiple lower-tail regimes, each with different empirical probabilities. This multiplicity is what enables informed choice—the algorithm selects among different upper-tail regimes based on which offers the highest likelihood of success. Joint-state memory is crucial because heavy‑tailed phenomena mean not all upper-tail regimes are equal. Some combinations of upper/lower partition moves across different dimensions lead to large improvements, whereas others do not.

At joint-state convergence points—where multiple Golden Section Search applications from different marginal distributions intersect—the algorithm evaluates which combination of upper or lower partition moves yields the highest expected improvement. These convergence points form a joint probability density field: regions of the hypercube with consistently high success rates are high‑density (favorable state configurations), while underperforming combinations are low‑density (unfavorable configurations). The algorithm preferentially samples high‑density regions, yet periodically verifies low‑density zones to detect regime shifts.

Crucially, each marginal distribution maintains its own state-transition history—a record of past transitions and their outcomes. In multidimensional space, marginal state frequencies form a probability metric space where distances between marginal distributions carry meaningful information. The Wasserstein distance (earth‑mover distance) measures the minimum "transport cost" required to reshape one probability distribution into another, making it ideal for comparing distances between marginal state distributions. When the Wasserstein distance between a marginal's current distribution and its baseline exceeds a threshold, the system flags a potential drift and triggers additional verification steps. This metric is particularly powerful in multidimensional settings because it captures the geometric structure of the probability metric space—how probability mass is distributed across the hypercube and how it shifts over time. In one‑dimensional settings, Wasserstein distance offers little advantage for understanding local probability structure, but in higher dimensions it becomes essential for tracking which joint state configurations are drifting and which remain stable.

Aharonov-Bohm Inspired Phase Dynamics

The algorithm incorporates insights from the Aharonov-Bohm effect in quantum mechanics, where charged particles acquire phase shifts from electromagnetic potentials even in field-free regions. This quantum phenomenon reveals that potentials—not just their gradients (fields)—have physical significance. We apply this insight to optimization:

Definition 7 (Momentum). The momentum vector μ ∈ ℝⁿ evolves according to: ⎧ α·μₖ + β·Δxₖ/(b-a), if f(xₖ₊₁) < f(xₖ) (improvement) μₖ₊₁ = ⎨ ⎩ γ·μₖ, otherwise (stagnation) where Δxₖ = xₖ₊₁ - xₖ, with parameters α = 0.8 (decay), β = 0.3 (acceleration), γ = 0.5. Definition 8 (Candidate Set). At iteration k, the candidate set 𝒦ₖ ⊂ Ω is: 𝒦ₖ = 𝒦ᶜʳᵒˢˢ ∪ 𝒦ᵇᵉˢᵗ ∪ 𝒦ᵐᵒᵐ ∪ 𝒦ʳⁿᵈ ∪ 𝒦ᵉˣᵗ where: 𝒦ᶜʳᵒˢˢ = {xₖ + σ·η·φ⁻¹·Δbₐ·eₐ : (i,j) ∈ ℰₙ, σ ∈ {-1,+1}, η ∈ {η₀, η₀/2, 2η₀}} 𝒦ᵇᵉˢᵗ = {xₖ + 0.2·φ⁻¹·(x* - xₖ)}, where x* = argmin_{t≤k} f(xₜ) 𝒦ᵐᵒᵐ = {xₖ + diag(b-a)·μₖ} 𝒦ʳⁿᵈ = {ξ}, ξ ~ Uniform(Ω) 𝒦ᵉˣᵗ = { clamp(gᵢ(ℂₖ), bounds) : gᵢ ∈ 𝒢 } with |𝒦ᶜʳᵒˢˢ| = 4·|ℰₙ|·3 = 12n(n-1) for n ≥ 2, adaptive step η₀ ∈ [0.02, 0.4]. External proposals gᵢ are normalized to the hyper-rectangle bounds by clamping each coordinate into [aᵢ, bᵢ] before frequency-state computation. Every candidate in 𝒦ₖ, including 𝒦ᵉˣᵗ, is passed through the same state key lookup, state probability rₛ, and phase-adjusted probability p' pipeline, with its source label preserved for debugging. Algorithm 1 (ε-Greedy MCDA Selection). Input: Candidate set 𝒦ₖ, phase-adjusted probabilities {p'(x)}_{x∈𝒦ₖ}, weights w, exploration rate ε Output: Selected candidate x* 1. Evaluate f(x) for all x ∈ 𝒦ₖ 2. Compute criteria: p'(x), Δf(x), step penalty, novelty, momentum alignment 3. Normalize criteria and score S(x) = Σᵢ wᵢ · cᵢ(x) 4. With probability ε: return x ~ Uniform(𝒦ₖ) 5. Return x* = argmax_{x∈𝒦ₖ} S(x) Theorem 2 (Almost Sure Convergence). Let {xₖ}_{k≥0} be the sequence generated by GoldenGradient on compact Ω ⊂ ℝⁿ with continuous f: Ω → ℝ. With random restarts every T iterations of stagnation, we have: P( lim inf f(xₖ) = f* ) = 1 k→∞ where f* = min_{x∈Ω} f(x). Proof sketch. The random restart mechanism ensures Ω is visited infinitely often. Combined with ε-greedy exploration, every neighborhood of f* is sampled with positive probability, and the greedy selection ensures improvements are retained. □

MCDA Weighting Defaults

Candidate selection uses a multi-criteria decision analysis (MCDA) score that blends state probability, predicted improvement, step size penalties, novelty, and momentum alignment. The defaults are chosen to balance exploitation (probability + improvement) with gentle exploration (novelty + momentum) while discouraging overly aggressive jumps.

Comprehensive Benchmark Suite

We evaluate Golden Gradient across five critical optimization dimensions, comparing it against state-of-the-art algorithms including CMA-ES, Nelder–Mead, Bayesian Optimization, Particle Swarm Optimization (PSO), and Differential Evolution (DE). Each dimension tests different aspects of optimization performance:

MCDA Weighting (Golden Gradient)

Tune how candidate ranking balances state probability, predicted improvement, step size penalties, novelty, and momentum alignment.

1D Base Layer

Evaluates algorithms in 1-dimensional space. This forms the base layer that will be cumulatively layered with additional dimensions in higher-dimensional topologies.

Performance Metrics

Detailed Results

Algorithm Best Value Mean Value Std Dev Convergence Rate Function Evals

Pareto Frontier (Golden Gradient)

# Position Objective Vector Scalar Score

2D Cumulative Topology

Evaluates algorithms in a 2-dimensional cumulative topology where each dimension layer contributes independently. Built from dimensions 1+2 layered together.

Performance Metrics

Detailed Results

Algorithm Best Value Mean Value Std Dev Success Rate Function Evals

Pareto Frontier (Golden Gradient)

# Position Objective Vector Scalar Score

3D Cumulative Topology

Evaluates algorithms in a 3-dimensional cumulative topology where each dimension layer contributes independently. Built from dimensions 1+2+3 layered together.

Performance Metrics

Scalability Analysis

Algorithm Final Value Time to Converge Memory Usage Efficiency Score

Pareto Frontier (Golden Gradient)

# Position Objective Vector Scalar Score

4D Cumulative Topology

Evaluates algorithms in a 4-dimensional cumulative topology where each dimension layer contributes independently. Built from dimensions 1+2+3+4 layered together.

Performance Metrics

Noise Robustness

Algorithm Best Value Mean Value Robustness Score Sample Efficiency

Pareto Frontier (Golden Gradient)

# Position Objective Vector Scalar Score

5D Cumulative Topology

Evaluates algorithms in a 5-dimensional cumulative topology where each dimension layer contributes independently. Built from dimensions 1+2+3+4+5 layered together.

Performance Metrics

Adaptation Analysis

Algorithm Tracking Error Recovery Time Drift Detection Adaptability Score

Pareto Frontier (Golden Gradient)

# Position Objective Vector Scalar Score

References

For background, see Kantorovich–Rubinstein duality (Wasserstein-1) and the Aharonov–Bohm effect.