Regularised p-Adic Linear Regression

An AI-Driven Mathematical Research Project

Project Overview

This research project investigates the effects of regularisation on p-adic linear regression. Building on the foundational work in arXiv:2510.00043v1, which established that an n-dimensional hyperplane minimizing p-adic sum of distances must pass through at least n+1 points, we explore what happens when regularisation is introduced.

Key Research Question

When we add regularisation to p-adic linear regression, how does the number of data points that the optimal hyperplane passes through change as a function of the regularisation strength?

Mathematical Background

p-Adic Valuation and Distance

For a prime p and rational number x = pk · (a/b) where gcd(p, ab) = 1:

Unregularised p-Adic Regression

The standard p-adic regression loss function is:

L(β) = Σi |yi − Xi·β|p = Σi p−vp(yi − Xi·β)

Regularised p-Adic Regression

We explore adding a regularisation term:

Lreg(β) = Σi |yi − Xi·β|p + λ · R(β)

where R(β) could be:

Current Status

Last Updated: 2025-12-06 (Day 39)

Current Focus: Formal Lean 4 proofs - second theorem fully proven!

Recent Progress (Day 39):

Previous Progress (Day 38):

Key Theorems Proven:

Next Steps: Complete kMin_implies_increase proof, formalise threshold formula.

Formal Proofs Status

TheoremStatusFile
R Decreases at TransitionsPROVENlean/PadicRegularisation/Basic.lean
Pareto Frontier MonotonicityPROVENlean/PadicRegularisation/Basic.lean
k_min PredictorStatement Formalizedlean/PadicRegularisation/Basic.lean