<?xml version="1.0" encoding="UTF-8"?>
<rss  xmlns:atom="http://www.w3.org/2005/Atom"
      xmlns:media="http://search.yahoo.com/mrss/"
      xmlns:content="http://purl.org/rss/1.0/modules/content/"
      xmlns:dc="http://purl.org/dc/elements/1.1/"
      version="2.0">
<channel>
<title>Yuda Bi — Research Notes</title>
<link>https://yuda-bi.com/notes.html</link>
<atom:link href="https://yuda-bi.com/notes.xml" rel="self" type="application/rss+xml"/>
<description>Technical notes on statistical physics, information geometry, spectral methods, and theoretical neuroscience.</description>
<language>en</language>
<generator>quarto-1.9.37</generator>
<lastBuildDate>Tue, 07 Apr 2026 00:00:00 GMT</lastBuildDate>
<item>
  <title>Spectral Methods in Brain Network Analysis</title>
  <dc:creator>Yuda Bi</dc:creator>
  <link>https://yuda-bi.com/notes/2026-04-spectral-brain-networks/</link>
  <description><![CDATA[




<section id="graph-laplacian-and-brain-networks" class="level2">
<h2 class="anchored" data-anchor-id="graph-laplacian-and-brain-networks">Graph Laplacian and Brain Networks</h2>
<p>Given a brain connectivity matrix <img src="https://latex.codecogs.com/png.latex?W%20%5Cin%20%5Cmathbb%7BR%7D%5E%7Bn%20%5Ctimes%20n%7D"> where <img src="https://latex.codecogs.com/png.latex?w_%7Bij%7D"> represents the functional connectivity between regions <img src="https://latex.codecogs.com/png.latex?i"> and <img src="https://latex.codecogs.com/png.latex?j">, the <strong>normalized graph Laplacian</strong> is:</p>
<p><img src="https://latex.codecogs.com/png.latex?%0A%5Cmathcal%7BL%7D%20=%20I%20-%20D%5E%7B-1/2%7D%20W%20D%5E%7B-1/2%7D%0A"></p>
<p>where <img src="https://latex.codecogs.com/png.latex?D%20=%20%5Ctext%7Bdiag%7D(d_1,%20%5Cldots,%20d_n)"> is the degree matrix with <img src="https://latex.codecogs.com/png.latex?d_i%20=%20%5Csum_j%20w_%7Bij%7D">.</p>
</section>
<section id="spectral-decomposition" class="level2">
<h2 class="anchored" data-anchor-id="spectral-decomposition">Spectral Decomposition</h2>
<p>The eigendecomposition of <img src="https://latex.codecogs.com/png.latex?%5Cmathcal%7BL%7D"> yields:</p>
<p><img src="https://latex.codecogs.com/png.latex?%0A%5Cmathcal%7BL%7D%20=%20U%20%5CLambda%20U%5ET%20=%20%5Csum_%7Bk=0%7D%5E%7Bn-1%7D%20%5Clambda_k%20%5Cmathbf%7Bu%7D_k%20%5Cmathbf%7Bu%7D_k%5ET%0A"></p>
<p>where <img src="https://latex.codecogs.com/png.latex?0%20=%20%5Clambda_0%20%5Cleq%20%5Clambda_1%20%5Cleq%20%5Ccdots%20%5Cleq%20%5Clambda_%7Bn-1%7D%20%5Cleq%202">.</p>
<section id="key-properties" class="level3">
<h3 class="anchored" data-anchor-id="key-properties">Key Properties</h3>
<ol type="1">
<li><strong>Number of zero eigenvalues</strong> = number of connected components</li>
<li>The <strong>Fiedler value</strong> <img src="https://latex.codecogs.com/png.latex?%5Clambda_1"> measures algebraic connectivity</li>
<li>The <strong>spectral gap</strong> <img src="https://latex.codecogs.com/png.latex?%5Clambda_1%20-%20%5Clambda_0"> indicates community separation strength</li>
</ol>
</section>
</section>
<section id="participation-ratio" class="level2">
<h2 class="anchored" data-anchor-id="participation-ratio">Participation Ratio</h2>
<p>The <strong>effective rank</strong> via participation ratio provides a model-free measure of spectral spread:</p>
<p><img src="https://latex.codecogs.com/png.latex?%0A%5Ctext%7BPR%7D(%5Cboldsymbol%7B%5Clambda%7D)%20=%20%5Cfrac%7B%5Cleft(%5Csum_k%20%5Clambda_k%5Cright)%5E2%7D%7B%5Csum_k%20%5Clambda_k%5E2%7D%0A"></p>
<p>This ranges from 1 (single dominant eigenvalue) to <img src="https://latex.codecogs.com/png.latex?n"> (uniform spectrum), capturing the effective dimensionality of the network.</p>
</section>
<section id="connection-to-statistical-physics" class="level2">
<h2 class="anchored" data-anchor-id="connection-to-statistical-physics">Connection to Statistical Physics</h2>
<p>The graph Laplacian connects to the partition function of a Gaussian field on the network:</p>
<p><img src="https://latex.codecogs.com/png.latex?%0AZ(%5Cbeta)%20=%20%5Cint%20%5Cexp%5Cleft(-%5Cfrac%7B%5Cbeta%7D%7B2%7D%20%5Cmathbf%7Bx%7D%5ET%20%5Cmathcal%7BL%7D%20%5Cmathbf%7Bx%7D%5Cright)%20d%5Cmathbf%7Bx%7D%20=%20%5Cprod_%7Bk=1%7D%5E%7Bn-1%7D%20%5Csqrt%7B%5Cfrac%7B2%5Cpi%7D%7B%5Cbeta%20%5Clambda_k%7D%7D%0A"></p>
<p>The free energy <img src="https://latex.codecogs.com/png.latex?F%20=%20-%5Cfrac%7B1%7D%7B%5Cbeta%7D%5Cln%20Z"> encodes thermodynamic properties of signal propagation on the network.</p>


</section>

 ]]></description>
  <category>spectral methods</category>
  <category>brain networks</category>
  <category>graph theory</category>
  <guid>https://yuda-bi.com/notes/2026-04-spectral-brain-networks/</guid>
  <pubDate>Tue, 07 Apr 2026 00:00:00 GMT</pubDate>
</item>
<item>
  <title>SpecRNA-QA: Why Spectral Graph Features See What Local Metrics Miss</title>
  <dc:creator>Yuda Bi</dc:creator>
  <link>https://yuda-bi.com/notes/2026-04-specrnaq/</link>
  <description><![CDATA[




<div class="quarto-figure quarto-figure-center">
<figure class="figure">
<p><img src="https://yuda-bi.com/notes/2026-04-specrnaq/featured.png" class="img-fluid figure-img" style="width:90.0%"></p>
<figcaption>RNA 3D structure quality visualization: C4’ deviation from best model (blue = 0 Å, red = 50 Å). Structures with correct local geometry but misplaced domains show large red regions — exactly the failure mode that spectral methods detect.</figcaption>
</figure>
</div>
<section id="the-problem-local-correctness-global-failure" class="level2">
<h2 class="anchored" data-anchor-id="the-problem-local-correctness-global-failure">The Problem: Local Correctness, Global Failure</h2>
<p>Predicting RNA 3D structure is one of the open frontiers of structural biology. Tools like AlphaFold3 and RoseTTAFold2NA now generate thousands of candidate structures — but <strong>how do you know which ones are right?</strong></p>
<p>Existing quality assessment (QA) methods evaluate <strong>local</strong> atomic contacts: bond angles, clash scores, pairwise distance distributions. These work well when errors are local. But they fail catastrophically in a common and important failure mode: <strong>the local structure is correct, but entire domains are misplaced.</strong> A helix can be perfectly folded yet docked into the wrong pocket. Every bond angle checks out; the global topology is wrong.</p>
<p>This is exactly the regime where RNA QA matters most — large, multi-domain structures where the combinatorial space of domain arrangements dwarfs the local conformational space.</p>
</section>
<section id="the-insight-global-topology-lives-in-the-spectrum" class="level2">
<h2 class="anchored" data-anchor-id="the-insight-global-topology-lives-in-the-spectrum">The Insight: Global Topology Lives in the Spectrum</h2>
<p>A 3D molecular structure is, at its core, a <strong>graph</strong>: nucleotides are nodes, spatial contacts are edges. The spectrum of the graph Laplacian — the eigenvalues of <img src="https://latex.codecogs.com/png.latex?%5Cmathcal%7BL%7D%20=%20I%20-%20D%5E%7B-1/2%7DWD%5E%7B-1/2%7D"> — encodes the global connectivity pattern in a way that is invariant to rotation, translation, and node relabeling.</p>
<p>The key mathematical facts:</p>
<ul>
<li>The <strong>number of zero eigenvalues</strong> equals the number of connected components</li>
<li>The <strong>Fiedler value</strong> <img src="https://latex.codecogs.com/png.latex?%5Clambda_1"> measures how easily the graph can be bisected — a proxy for global compactness</li>
<li>The <strong>spectral gap</strong> <img src="https://latex.codecogs.com/png.latex?%5Clambda_1%20-%20%5Clambda_0"> quantifies community separation</li>
<li><strong>Heat-kernel traces</strong> <img src="https://latex.codecogs.com/png.latex?Z(t)%20=%20%5Csum_k%20e%5E%7B-%5Clambda_k%20t%7D"> capture multi-scale diffusion: small <img src="https://latex.codecogs.com/png.latex?t"> probes local geometry, large <img src="https://latex.codecogs.com/png.latex?t"> probes global topology</li>
<li>The <strong>participation ratio</strong> <img src="https://latex.codecogs.com/png.latex?%5Cmathrm%7BPR%7D(%5Cboldsymbol%7B%5Clambda%7D)%20=%20(%5Csum_k%20%5Clambda_k)%5E2%20/%20%5Csum_k%20%5Clambda_k%5E2"> measures effective spectral dimensionality</li>
</ul>
<p>A misplaced domain changes the large-<img src="https://latex.codecogs.com/png.latex?t"> heat-kernel trace (disrupted long-range diffusion) while leaving the small-<img src="https://latex.codecogs.com/png.latex?t"> trace nearly intact (local contacts are fine). This is precisely the information that local metrics cannot access.</p>
</section>
<section id="the-method" class="level2">
<h2 class="anchored" data-anchor-id="the-method">The Method</h2>
<p><strong>SpecRNA-QA</strong> builds on this insight with a practical pipeline:</p>
<ol type="1">
<li><p><strong>Multi-scale contact graphs</strong>: Construct contact networks at multiple distance thresholds (8Å, 10Å, 12Å, 15Å), capturing different spatial resolutions of the RNA architecture.</p></li>
<li><p><strong>Spectral feature extraction</strong>: From each graph’s normalized Laplacian, extract ~312 features:</p>
<ul>
<li>Eigenvalue statistics (mean, variance, skewness, kurtosis of <img src="https://latex.codecogs.com/png.latex?%5C%7B%5Clambda_k%5C%7D">)</li>
<li>Heat-kernel traces at multiple diffusion times <img src="https://latex.codecogs.com/png.latex?Z(t)"> for <img src="https://latex.codecogs.com/png.latex?t%20%5Cin%20%5C%7B0.1,%200.5,%201,%202,%205,%2010%5C%7D"></li>
<li>Participation ratios and effective rank measures</li>
<li>Spectral gap and algebraic connectivity</li>
<li>Normalized Laplacian entropy <img src="https://latex.codecogs.com/png.latex?H%20=%20-%5Csum_k%20%5Chat%7B%5Clambda%7D_k%20%5Clog%20%5Chat%7B%5Clambda%7D_k"></li>
</ul></li>
<li><p><strong>Learning-to-rank</strong>: An XGBRanker model trained to rank structures by quality within each target, using spectral features as input.</p></li>
</ol>
<p>The entire pipeline runs on <strong>CPU</strong> — no GPU required. Processing time: <strong>15 ms</strong> for a 100-nucleotide structure, <strong>~4.2 seconds</strong> for 800 nucleotides.</p>
</section>
<section id="results" class="level2">
<h2 class="anchored" data-anchor-id="results">Results</h2>
<section id="casp16-benchmark" class="level3">
<h3 class="anchored" data-anchor-id="casp16-benchmark">CASP16 Benchmark</h3>
<p>On the CASP16 RNA structure prediction assessment (42 targets, 7,368 models):</p>
<table class="caption-top table">
<thead>
<tr class="header">
<th>Method</th>
<th>Median Spearman <img src="https://latex.codecogs.com/png.latex?%5Crho"></th>
<th><img src="https://latex.codecogs.com/png.latex?p">-value vs.&nbsp;SpecRNA-QA</th>
</tr>
</thead>
<tbody>
<tr class="odd">
<td><strong>SpecRNA-QA (supervised)</strong></td>
<td><strong>0.689</strong></td>
<td>—</td>
</tr>
<tr class="even">
<td>Geometry baselines</td>
<td>0.465</td>
<td><img src="https://latex.codecogs.com/png.latex?1.2%20%5Ctimes%2010%5E%7B-10%7D"></td>
</tr>
</tbody>
</table>
</section>
<section id="where-it-matters-most-large-rnas" class="level3">
<h3 class="anchored" data-anchor-id="where-it-matters-most-large-rnas">Where It Matters Most: Large RNAs</h3>
<p>The advantage is most pronounced for large RNA structures (&gt;200 nucleotides), where the performance gap reaches <strong>+0.233</strong> in Spearman correlation. This is the regime where domain-level misplacements dominate — and where spectral features shine.</p>
<p>For small RNAs (&lt;100 nt), local metrics are often sufficient because there are few domains to misplace. The spectral advantage grows with structural complexity, exactly as the theory predicts.</p>
</section>
<section id="most-discriminative-features" class="level3">
<h3 class="anchored" data-anchor-id="most-discriminative-features">Most Discriminative Features</h3>
<p>Feature importance analysis reveals that the top-ranked features are <strong>heat-kernel traces at intermediate-to-large diffusion times</strong> — precisely the features that probe multi-scale and global transport geometry on the contact network. Local eigenvalue statistics (which probe small-scale structure) rank lower.</p>
<p>This confirms the theoretical motivation: the spectral approach works because it accesses the global information that local methods cannot reach.</p>
</section>
</section>
<section id="connection-to-the-broader-spectral-program" class="level2">
<h2 class="anchored" data-anchor-id="connection-to-the-broader-spectral-program">Connection to the Broader Spectral Program</h2>
<p>SpecRNA-QA is part of a broader research program applying spectral graph theory to structural biology:</p>
<ul>
<li><strong>SpecRNA-QA</strong> (RNA): Multi-scale Laplacian spectra for RNA 3D quality assessment → under review at <em>Briefings in Bioinformatics</em></li>
<li><strong>Spectral Coherence Index</strong> (Proteins): Participation-ratio effective rank of inter-model distance-variance matrices for protein ensemble QA, achieving AUC-ROC 0.973 on 110 NMR ensembles → under review at <em>IEEE JBHI</em> (<a href="https://arxiv.org/abs/2603.25880">arXiv:2603.25880</a>)</li>
</ul>
<p>Both methods share a design principle: <strong>model-free spectral features that are invariant to coordinate systems and capture global structural properties that local metrics miss.</strong></p>
</section>
<section id="try-it" class="level2">
<h2 class="anchored" data-anchor-id="try-it">Try It</h2>
<p>SpecRNA-QA is open source and easy to use:</p>
<div class="code-copy-outer-scaffold"><div class="sourceCode" id="cb1" style="background: #f1f3f5;"><pre class="sourceCode bash code-with-copy"><code class="sourceCode bash"><span id="cb1-1"><span class="fu" style="color: #4758AB;
background-color: null;
font-style: inherit;">git</span> clone https://github.com/yudabitrends/specrnaq</span>
<span id="cb1-2"><span class="bu" style="color: null;
background-color: null;
font-style: inherit;">cd</span> specrnaq</span>
<span id="cb1-3"><span class="ex" style="color: null;
background-color: null;
font-style: inherit;">pip</span> install <span class="at" style="color: #657422;
background-color: null;
font-style: inherit;">-e</span> .</span>
<span id="cb1-4"><span class="ex" style="color: null;
background-color: null;
font-style: inherit;">specrnaq</span> predict <span class="at" style="color: #657422;
background-color: null;
font-style: inherit;">--input</span> structures/ <span class="at" style="color: #657422;
background-color: null;
font-style: inherit;">--output</span> scores.csv</span></code></pre></div></div>
<p>Python 3.10+, CPU-only, no external dependencies beyond standard scientific Python.</p>
<hr>
</section>
<section id="papers" class="level2">
<h2 class="anchored" data-anchor-id="papers">Papers</h2>
<ol type="1">
<li><p>Ying Zhu, Huaiwen Zhang, Vince D. Calhoun<sup>†</sup>, <strong>Yuda Bi</strong><sup>†</sup>. <em>Spectral Graph Features Capture Global Topology for Reference-free RNA 3D Structure Quality Assessment.</em> Under review at Briefings in Bioinformatics.</p></li>
<li><p><strong>Yuda Bi</strong>, Huaiwen Zhang, Jingnan Sun, Vince D. Calhoun. <em>Spectral Coherence Index: A Model-Free Metric for Protein Structural Ensemble Quality Assessment.</em> Under review at IEEE JBHI. <a href="https://arxiv.org/abs/2603.25880">arXiv:2603.25880</a></p></li>
</ol>


</section>

 ]]></description>
  <category>spectral methods</category>
  <category>RNA structure</category>
  <category>bioinformatics</category>
  <category>graph theory</category>
  <guid>https://yuda-bi.com/notes/2026-04-specrnaq/</guid>
  <pubDate>Tue, 07 Apr 2026 00:00:00 GMT</pubDate>
  <media:content url="https://yuda-bi.com/notes/2026-04-specrnaq/featured.png" medium="image" type="image/png" height="111" width="144"/>
</item>
<item>
  <title>The Geometry of Invisible Forces: A Quartic Detection Theory</title>
  <dc:creator>Yuda Bi</dc:creator>
  <link>https://yuda-bi.com/notes/2026-04-quartic-detection-theory/</link>
  <description><![CDATA[




<section id="the-problem-forces-you-cannot-see" class="level2">
<h2 class="anchored" data-anchor-id="the-problem-forces-you-cannot-see">The Problem: Forces You Cannot See</h2>
<p>Hidden degrees of freedom are everywhere. The ocean drives climate through modes no single weather station resolves. Latent neural populations shape the signals recorded by any one electrode. Slow institutional forces move markets through channels invisible to any single asset’s price history. The central question is deceptively simple: <strong>given noisy observations, can you tell whether a hidden force is present?</strong></p>
<p>Classical statistics offers a reassuring answer — collect enough data and any nonzero signal emerges. Fisher information scales as <img src="https://latex.codecogs.com/png.latex?N"> (the sample size), so the detection boundary shrinks as <img src="https://latex.codecogs.com/png.latex?N%5E%7B-1/2%7D">. For a coupling strength <img src="https://latex.codecogs.com/png.latex?%5Clambda">, you need <img src="https://latex.codecogs.com/png.latex?N%20%5Csim%20%5Clambda%5E%7B-2%7D"> observations. Reasonable. Manageable.</p>
<p><strong>This answer is wrong.</strong> Not slightly wrong — wrong by orders of magnitude.</p>
</section>
<section id="the-quartic-law-why-detection-is-exponentially-harder" class="level2">
<h2 class="anchored" data-anchor-id="the-quartic-law-why-detection-is-exponentially-harder">The Quartic Law: Why Detection Is Exponentially Harder</h2>
<p>Consider the simplest possible hidden-variable problem. An observed time series <img src="https://latex.codecogs.com/png.latex?X_t"> obeys</p>
<p><img src="https://latex.codecogs.com/png.latex?%0AX_%7Bt+1%7D%20=%20a%5C,%20X_t%20+%20%5Clambda%5C,%20F_t%20+%20%5Cvarepsilon_t%0A"></p>
<p>where <img src="https://latex.codecogs.com/png.latex?F_t"> is a hidden persistent driver (<img src="https://latex.codecogs.com/png.latex?F_%7Bt+1%7D%20=%20b%5C,%20F_t%20+%20%5Ceta_t">) and <img src="https://latex.codecogs.com/png.latex?%5Cvarepsilon_t"> is noise. The coupling <img src="https://latex.codecogs.com/png.latex?%5Clambda"> controls how strongly the hidden world leaks into the visible one.</p>
<p>The power spectrum of <img src="https://latex.codecogs.com/png.latex?X_t"> is indeed perturbed at <img src="https://latex.codecogs.com/png.latex?O(%5Clambda%5E2)">. So far, so classical. But here is the catch: <strong>when you refit a reduced model (one without the hidden variable), the best-fit parameters shift to absorb most of that perturbation.</strong> The refitted model “explains away” the hidden signal by reparametrizing itself.</p>
<p>What survives is not the <img src="https://latex.codecogs.com/png.latex?O(%5Clambda%5E2)"> perturbation, but only its <strong>normal component</strong> — the part that cannot be absorbed by any reparametrization of the reduced model. And this residual is <img src="https://latex.codecogs.com/png.latex?O(%5Clambda%5E4)">.</p>
<p>The result is the <strong>quartic detection law</strong>:</p>
<p><img src="https://latex.codecogs.com/png.latex?%0AD_%7B%5Cmathrm%7BKL%7D%7D%5E%7B%5Cmin%7D(%5Clambda)%20=%20C%5C,%5Clambda%5E4%20+%20O(%5Clambda%5E6)%0A"></p>
<p>where <img src="https://latex.codecogs.com/png.latex?C"> is a system-specific constant and <img src="https://latex.codecogs.com/png.latex?D_%7B%5Cmathrm%7BKL%7D%7D%5E%7B%5Cmin%7D"> is the minimum Kullback-Leibler divergence between truth and the best-fit reduced model. The detection boundary becomes</p>
<p><img src="https://latex.codecogs.com/png.latex?%0A%5Cboxed%7B%5Clambda_c(N)%20%5C;%5Cpropto%5C;%20%5Cleft(%5Cfrac%7B%5Clog%20N%7D%7BN%7D%5Cright)%5E%7B1/4%7D%7D%0A"></p>
<p>For a 10% coupling (<img src="https://latex.codecogs.com/png.latex?%5Clambda/%5Csigma%20=%200.1">), the quartic penalty demands <img src="https://latex.codecogs.com/png.latex?%5Csim%2010%5E3"> times more data than the naive quadratic expectation. This is not a small correction — it is a qualitative change in what is experimentally feasible.</p>
</section>
<section id="the-geometry-tangent-absorption-on-statistical-manifolds" class="level2">
<h2 class="anchored" data-anchor-id="the-geometry-tangent-absorption-on-statistical-manifolds">The Geometry: Tangent Absorption on Statistical Manifolds</h2>
<p>The quartic law is not an accident of the specific model. It is a <strong>geometric theorem on statistical manifolds</strong>.</p>
<p>Let <img src="https://latex.codecogs.com/png.latex?%5Cmathcal%7BM%7D_0"> be the manifold of reduced (null) models, parametrized by <img src="https://latex.codecogs.com/png.latex?%5Cboldsymbol%7B%5Ctheta%7D">. The true distribution under hidden forcing is</p>
<p><img src="https://latex.codecogs.com/png.latex?%0Ap_%7B%5Cmathrm%7Btrue%7D%7D(x%20%5Cmid%20%5Clambda)%20=%20p_0(x%20%5Cmid%20%5Cboldsymbol%7B%5Ctheta%7D_0)%5C,%5Cbigl(1%20+%20%5Clambda%5E2%5C,%20h(x)%20+%20O(%5Clambda%5E4)%5Cbigr)%0A"></p>
<p>The perturbation <img src="https://latex.codecogs.com/png.latex?h"> has a unique decomposition into tangent and normal components relative to <img src="https://latex.codecogs.com/png.latex?%5Cmathcal%7BM%7D_0">:</p>
<p><img src="https://latex.codecogs.com/png.latex?%0Ah%20=%20%5Cunderbrace%7B%5CPi_T%5C,%20h%7D_%7B%5Ctext%7Babsorbed%20by%20refitting%7D%7D%20+%20%5Cunderbrace%7BR%7D_%7B%5Ctext%7Bdetectable%20residual%7D%7D%0A"></p>
<p>where <img src="https://latex.codecogs.com/png.latex?%5CPi_T"> is the <img src="https://latex.codecogs.com/png.latex?L%5E2(p_0)">-projection onto the tangent space <img src="https://latex.codecogs.com/png.latex?T_%7B%5Cboldsymbol%7B%5Ctheta%7D_0%7D%5Cmathcal%7BM%7D_0">. The tangent part is indistinguishable from a parameter shift; only the normal residual <img src="https://latex.codecogs.com/png.latex?R%20=%20(I%20-%20%5CPi_T)%5C,h"> is genuinely detectable. Since <img src="https://latex.codecogs.com/png.latex?h"> enters at <img src="https://latex.codecogs.com/png.latex?O(%5Clambda%5E2)">, the squared residual — hence the KL divergence — is <img src="https://latex.codecogs.com/png.latex?O(%5Clambda%5E4)">.</p>
<p>This is Efron’s statistical curvature repurposed for detection: the “curvature” of the embedding of truth relative to the null manifold controls how much signal survives refitting.</p>
</section>
<section id="the-pairing-principle-one-probe-is-blind" class="level2">
<h2 class="anchored" data-anchor-id="the-pairing-principle-one-probe-is-blind">The Pairing Principle: One Probe Is Blind</h2>
<p>The geometry yields a stark operational consequence. Consider <img src="https://latex.codecogs.com/png.latex?n"> sensors sharing a common hidden source:</p>
<p><img src="https://latex.codecogs.com/png.latex?%0AY_%7Bij%7D%20=%20%5Cmu%20+%20%5Clambda%5C,%20u_i%20+%20%5Cvarepsilon_%7Bij%7D,%20%5Cqquad%20u_i%20%5Csim%20N(0,1)%0A"></p>
<p>The quartic coefficient is</p>
<p><img src="https://latex.codecogs.com/png.latex?%0AC%20=%20%5Cfrac%7Bn(n-1)%7D%7B4%5Csigma%5E4%7D%0A"></p>
<p>The combinatorial factor <img src="https://latex.codecogs.com/png.latex?n(n-1)%20=%202%5Cbinom%7Bn%7D%7B2%7D"> counts <strong>sensor pairs</strong> sharing the hidden effect. When <img src="https://latex.codecogs.com/png.latex?n%20=%201">, <img src="https://latex.codecogs.com/png.latex?C%20=%200"> identically — no amount of data can detect the hidden variable from a single probe. This is not a power problem; it is a <strong>structural impossibility</strong>.</p>
<p>The <strong>pairing principle</strong> states:</p>
<blockquote class="blockquote">
<p>Detecting a hidden degree of freedom requires at least two probes sharing the hidden effect. Detection power grows as the number of probe pairs <img src="https://latex.codecogs.com/png.latex?%5Cbinom%7Bn%7D%7B2%7D">, with diminishing returns: <img src="https://latex.codecogs.com/png.latex?%5Clambda_c%20%5Cpropto%20n%5E%7B-1/2%7D">, not <img src="https://latex.codecogs.com/png.latex?n%5E%7B-1%7D">.</p>
</blockquote>
<p>For Gaussian observations, this is exact. A single electrode in a neural recording, a single climate station, a single asset price — each is <strong>provably blind</strong> to a latent common driver, regardless of record length.</p>
</section>
<section id="the-dark-regime-when-timescales-coalesce" class="level2">
<h2 class="anchored" data-anchor-id="the-dark-regime-when-timescales-coalesce">The Dark Regime: When Timescales Coalesce</h2>
<p>Even with the quartic law in hand, there is a deeper obstruction. In the spectral setting, the quartic coefficient takes the closed form</p>
<p><img src="https://latex.codecogs.com/png.latex?%0AC(a,%20b)%20=%20%5Cfrac%7B%5Csigma_%5Ceta%5E4%5C,%20b%5E2%5C,%20(a%20-%20b)%5E2%7D%7B2%5C,%5Csigma_%5Cvarepsilon%5E4%5C,(1%20-%20b%5E2)%5E3%5C,(1%20-%20ab)%5E2%7D%0A"></p>
<p>This coefficient <strong>vanishes</strong> when <img src="https://latex.codecogs.com/png.latex?a%20=%20b"> — when the intrinsic relaxation timescale of the observed system exactly matches that of the hidden driver. At this <strong>timescale coalescence</strong>, the hidden driver’s spectral signature is perfectly tangent to the null manifold. The hidden forcing becomes <em>spectrally dark</em>: present, dynamically active, yet locally invisible to any single-channel spectral test.</p>
<p>The data cost diverges as the dark boundary is approached:</p>
<p><img src="https://latex.codecogs.com/png.latex?%0A%5Clambda_c%20%5C;%5Cpropto%5C;%20%7Ca%20-%20b%7C%5E%7B-1/2%7D%20%5Cto%20%5Cinfty%20%5Cquad%20%5Ctext%7Bas%20%7D%20a%20%5Cto%20b%0A"></p>
<p>This is not merely a detection difficulty — it is a geometric singularity of the inference problem itself.</p>
</section>
<section id="breaking-the-impossibility-the-cross-spectral-escape" class="level2">
<h2 class="anchored" data-anchor-id="breaking-the-impossibility-the-cross-spectral-escape">Breaking the Impossibility: The Cross-Spectral Escape</h2>
<p>The single-channel impossibility is not the end of the story. It is the beginning of a deeper one.</p>
<p>Lucente et al.&nbsp;proved that no time-irreversibility measure can detect departure from equilibrium in a scalar Gaussian time series from a linear system. This seems like a fundamental wall. But it has a geometric loophole: <strong>it applies only to diagonal (single-channel) observations.</strong></p>
<p>When a second channel shares the same hidden driver, the <strong>cross spectrum</strong> — the off-diagonal block of the spectral matrix — provides a detection channel that is orthogonal to the entire diagonal null manifold. The cross-spectral contribution obeys its own quartic law:</p>
<p><img src="https://latex.codecogs.com/png.latex?%0AD_%7B%5Cmathrm%7Bcross%7D%7D(%5Clambda)%20=%20C_%7B%5Cmathrm%7Bcross%7D%7D%5C,%5Clambda%5E4%20+%20O(%5Clambda%5E6)%0A"></p>
<p>with a remarkable property: the coefficient <img src="https://latex.codecogs.com/png.latex?C_%7B%5Cmathrm%7Bcross%7D%7D"> is <strong>exactly independent of the observed-channel dynamics</strong>. All dependence on the transfer functions <img src="https://latex.codecogs.com/png.latex?H_1,%20H_2"> of the observed channels cancels identically. The cross-spectral detectability is determined solely by the hidden mode’s spectral density:</p>
<p><img src="https://latex.codecogs.com/png.latex?%0AC_%7B%5Cmathrm%7Bcross%7D%7D%20=%20%5Cfrac%7Bu_1%5E2%5C,%20u_2%5E2%7D%7B%5Csigma_%7B%5Cvarepsilon_1%7D%5E2%5C,%20%5Csigma_%7B%5Cvarepsilon_2%7D%5E2%7D%20%5Ccdot%20I_F,%20%5Cqquad%20I_F%20=%20%5Cfrac%7B1%7D%7B4%5Cpi%7D%5Cint_%7B-%5Cpi%7D%5E%7B%5Cpi%7D%20S_F(%5Comega)%5E2%5C,%20d%5Comega%0A"></p>
<p>Crucially, <img src="https://latex.codecogs.com/png.latex?C_%7B%5Cmathrm%7Bcross%7D%7D%20%3E%200"> <strong>at exact timescale coalescence</strong>, where all single-channel measures vanish. The cross spectrum breaks the dark regime.</p>
</section>
<section id="the-thermodynamic-bridge-detectability-certifies-irreversibility" class="level2">
<h2 class="anchored" data-anchor-id="the-thermodynamic-bridge-detectability-certifies-irreversibility">The Thermodynamic Bridge: Detectability Certifies Irreversibility</h2>
<p>The connection deepens at the level of thermodynamics. For a one-way coupled Ornstein-Uhlenbeck system, the full-system entropy production rate (EPR) is exactly</p>
<p><img src="https://latex.codecogs.com/png.latex?%0A%5Cdot%7B%5CSigma%7D_%7B%5Cmathrm%7Btotal%7D%7D%20=%20%5Calpha_2%5C,%5Clambda%5E2%0A"></p>
<p>and the EPR-detectability bridge reads</p>
<p><img src="https://latex.codecogs.com/png.latex?%0A%5Cdot%7B%5CSigma%7D_%7B%5Cmathrm%7Btotal%7D%7D%5E2%20=%20%5Cfrac%7B%5Calpha_2%5E2%7D%7BC_%7B%5Cmathrm%7Bcross%7D%7D%7D%5C,%20D_%7B%5Cmathrm%7Bcross%7D%7D%20+%20O(%5Clambda%5E6)%0A"></p>
<p>This means: <strong>if the cross-spectral divergence is positive, the full system is certifiably out of equilibrium.</strong> Cross-spectral structure witnesses entropy production even when every single-channel estimator of time-irreversibility returns zero.</p>
<p>A single probe can sit inside a system with arbitrarily large true entropy production and measure <em>exactly</em> zero irreversibility. Two probes, sharing the same hidden bath, break this thermodynamic blindness.</p>
</section>
<section id="the-unified-picture" class="level2">
<h2 class="anchored" data-anchor-id="the-unified-picture">The Unified Picture</h2>
<p>The three results assemble into a coherent geometric theory of hidden-variable detectability:</p>
<table class="caption-top table">
<colgroup>
<col style="width: 25%">
<col style="width: 29%">
<col style="width: 44%">
</colgroup>
<thead>
<tr class="header">
<th>Layer</th>
<th>Result</th>
<th>Key Object</th>
</tr>
</thead>
<tbody>
<tr class="odd">
<td><strong>Universal law</strong></td>
<td>Quartic onset <img src="https://latex.codecogs.com/png.latex?D%20%5Csim%20%5Clambda%5E4"></td>
<td>Normal residual <img src="https://latex.codecogs.com/png.latex?R%20=%20(I%20-%20%5CPi_T)%5C,h"></td>
</tr>
<tr class="even">
<td><strong>Structural impossibility</strong></td>
<td>Single probe is blind (<img src="https://latex.codecogs.com/png.latex?C%20=%200"> at <img src="https://latex.codecogs.com/png.latex?n=1">)</td>
<td>Pairing principle: need <img src="https://latex.codecogs.com/png.latex?%5Cgeq%202"> probes</td>
</tr>
<tr class="odd">
<td><strong>Spectral darkness</strong></td>
<td><img src="https://latex.codecogs.com/png.latex?C%20%5Cto%200"> at timescale coalescence</td>
<td>Tangent alignment of hidden spectrum</td>
</tr>
<tr class="even">
<td><strong>Cross-spectral escape</strong></td>
<td><img src="https://latex.codecogs.com/png.latex?C_%7B%5Cmathrm%7Bcross%7D%7D%20%3E%200"> at coalescence</td>
<td>Off-diagonal orthogonality</td>
</tr>
<tr class="odd">
<td><strong>Thermodynamic certification</strong></td>
<td><img src="https://latex.codecogs.com/png.latex?D_%7B%5Cmathrm%7Bcross%7D%7D%20%3E%200%20%5CRightarrow%20%5Cdot%7B%5CSigma%7D%20%3E%200"></td>
<td>EPR-detectability bridge</td>
</tr>
</tbody>
</table>
<p>The hierarchy is:</p>
<ol type="1">
<li>A single observation channel is <strong>exactly dark</strong> — hidden forcing is indistinguishable from equilibrium, for all coupling strengths, at all sample sizes.</li>
<li>Two channels sharing the hidden driver restore quartic detectability via the cross spectrum, even at the coalescence singularity where auto-spectral methods fail.</li>
<li>The cross-spectral witness certifies that the full system is out of equilibrium, linking observable spectral structure to the thermodynamic arrow of time.</li>
</ol>
</section>
<section id="implications" class="level2">
<h2 class="anchored" data-anchor-id="implications">Implications</h2>
<p><strong>For neuroscience</strong>: Detecting latent neural inputs requires multi-channel recording. The pairing principle prescribes exactly how many electrodes are needed given estimates of coupling and noise. A single fMRI voxel or EEG electrode is provably blind to a shared latent source.</p>
<p><strong>For climate science</strong>: Detecting unresolved ocean forcing in climate records requires spatially distributed stations. A single station, no matter how long the record, cannot distinguish forced variability from intrinsic noise. The dark regime at timescale coalescence explains why slow ocean modes are so difficult to identify.</p>
<p><strong>For stochastic thermodynamics</strong>: The single-channel impossibility is not a limitation of current methods — it is a geometric fact about the statistical manifold. The cross-spectral escape provides a constructive route to certifying irreversibility from partial observations.</p>
<p><strong>For experimental design</strong>: The quartic law transforms the question “how much data do I need?” into “how many probes do I need, and where?” The answer: at least two, sharing the hidden effect, with detection power growing as the number of probe pairs.</p>
<hr>
</section>
<section id="papers-in-this-series" class="level2">
<h2 class="anchored" data-anchor-id="papers-in-this-series">Papers in This Series</h2>
<ol type="1">
<li><p><strong>Yuda Bi</strong>, Vince D. Calhoun. <em>Why Single Probes Cannot Detect Hidden Forcing: A Quartic Detection Law.</em> Under review at Physical Review Letters.</p></li>
<li><p><strong>Yuda Bi</strong>, Vince D. Calhoun. <em>Cross Spectra Break the Single-Channel Impossibility.</em> Under review at Physical Review Letters.</p></li>
<li><p><strong>Yuda Bi</strong>, Chenyu Zhang, Vince D. Calhoun. <em>Timescale Coalescence Makes Hidden Persistent Forcing Spectrally Dark.</em> Under review at Physical Review E. <a href="https://arxiv.org/abs/2603.20917">arXiv:2603.20917</a></p></li>
<li><p><strong>Yuda Bi</strong>, Vince D. Calhoun. <em>Conditioning on a Volatility Proxy Compresses the Apparent Timescale of Collective Market Correlation.</em> Under review at Physical Review E. <a href="https://arxiv.org/abs/2603.14072">arXiv:2603.14072</a></p></li>
</ol>


</section>

 ]]></description>
  <category>statistical physics</category>
  <category>information geometry</category>
  <category>hidden variables</category>
  <category>spectral methods</category>
  <guid>https://yuda-bi.com/notes/2026-04-quartic-detection-theory/</guid>
  <pubDate>Tue, 07 Apr 2026 00:00:00 GMT</pubDate>
</item>
<item>
  <title>Structure Sets the Stage: How Gray Matter Geometry Constrains Functional Brain Organization</title>
  <dc:creator>Yuda Bi</dc:creator>
  <link>https://yuda-bi.com/notes/2026-04-sfcoupling/</link>
  <description><![CDATA[




<div class="quarto-figure quarto-figure-center">
<figure class="figure">
<p><img src="https://yuda-bi.com/notes/2026-04-sfcoupling/featured.png" class="img-fluid figure-img" style="width:100.0%"></p>
<figcaption><strong>Spectral fingerprint of structure-function coupling.</strong> (a) Singular value spectrum of the GM→FNC coefficient matrix showing rapid spectral decay — the coupling is low-rank. (b) Principal angle cosines between GM-selected and FNC-dominant subspaces: near-perfect alignment in ~3 dimensions, then rapid falloff. (c) The signature gap: subspace overlap reaches 0.45 while variance explained (<img src="https://latex.codecogs.com/png.latex?R%5E2">) stays at ~0.06.</figcaption>
</figure>
</div>
<section id="the-paradox-weak-prediction-strong-geometry" class="level2">
<h2 class="anchored" data-anchor-id="the-paradox-weak-prediction-strong-geometry">The Paradox: Weak Prediction, Strong Geometry</h2>
<p>A persistent puzzle in systems neuroscience: gray matter morphometry (regional volume, thickness, surface area) predicts functional network connectivity (FNC) poorly — typically explaining only a few percent of variance. This has led many to dismiss the structure-function relationship as negligible for gray matter, focusing attention on white-matter tractography instead.</p>
<p>But <strong>variance explained is the wrong metric</strong>. It measures whether structure predicts the <em>amplitude</em> of functional variation — how strongly each person’s connectivity deviates from the mean. A more fundamental question is whether structure predicts the <em>directions</em> — the coordinate axes along which functional variation occurs at all.</p>
<p>These are mathematically distinct quantities. A linear map <img src="https://latex.codecogs.com/png.latex?%5Chat%7BY%7D%20=%20XB"> from GM features (<img src="https://latex.codecogs.com/png.latex?X%20%5Cin%20%5Cmathbb%7BR%7D%5E%7BN%20%5Ctimes%20p%7D">) to FNC (<img src="https://latex.codecogs.com/png.latex?Y%20%5Cin%20%5Cmathbb%7BR%7D%5E%7BN%20%5Ctimes%20q%7D">) can have low <img src="https://latex.codecogs.com/png.latex?R%5E2"> (weak amplitude prediction) while its coefficient matrix <img src="https://latex.codecogs.com/png.latex?B"> has column space closely aligned with the dominant subspace of <img src="https://latex.codecogs.com/png.latex?Y"> (strong directional alignment).</p>
</section>
<section id="the-method-nuclear-norm-regularization-and-svd" class="level2">
<h2 class="anchored" data-anchor-id="the-method-nuclear-norm-regularization-and-svd">The Method: Nuclear Norm Regularization and SVD</h2>
<p>To disentangle amplitude from geometry, we use a spectral regularization framework:</p>
<p><img src="https://latex.codecogs.com/png.latex?%0A%5Cmin_B%20%5C;%20%5Cfrac%7B1%7D%7B2N%7D%5C%7CY%20-%20XB%5C%7C_F%5E2%20+%20%5Clambda%20%5C%7CB%5C%7C_*%0A"></p>
<p>where <img src="https://latex.codecogs.com/png.latex?%5C%7CB%5C%7C_*%20=%20%5Csum_i%20%5Csigma_i(B)"> is the <strong>nuclear norm</strong> — the tightest convex relaxation of matrix rank. This encourages the learned mapping to be low-rank, concentrating the structure-function relationship into a small number of principal modes.</p>
<p>The proximal step has a closed-form solution via soft singular value thresholding:</p>
<p><img src="https://latex.codecogs.com/png.latex?%0AB%5E%7B(t+1)%7D%20=%20U%5C,%5Cmathrm%7Bdiag%7D%5Cbigl((%5Csigma_i%20-%20%5Clambda/L)_+%5Cbigr)%5C,V%5E%5Ctop%0A"></p>
<p>The SVD of the resulting coefficient matrix <img src="https://latex.codecogs.com/png.latex?B%20=%20U%5CSigma%20V%5E%5Ctop"> gives us:</p>
<ul>
<li><img src="https://latex.codecogs.com/png.latex?V">: the <strong>FNC directions</strong> most aligned with GM variation (the “stage geometry”)</li>
<li><img src="https://latex.codecogs.com/png.latex?U">: the <strong>GM features</strong> driving each coupling mode (the “structural scaffolding”)</li>
<li><img src="https://latex.codecogs.com/png.latex?%5CSigma">: the <strong>coupling strength</strong> per mode (spectral concentration)</li>
</ul>
</section>
<section id="measuring-directional-alignment-subspace-overlap" class="level2">
<h2 class="anchored" data-anchor-id="measuring-directional-alignment-subspace-overlap">Measuring Directional Alignment: Subspace Overlap</h2>
<p>To compare GM-selected functional directions with the dominant directions of FNC variation itself, we compute <strong>principal angles</strong> between two subspaces:</p>
<p><img src="https://latex.codecogs.com/png.latex?%0A%5Ccos%5Ctheta_i%20=%20%5Csigma_i(V_1%5E%5Ctop%20V_2)%0A"></p>
<p>where <img src="https://latex.codecogs.com/png.latex?V_1"> are the top-<img src="https://latex.codecogs.com/png.latex?k"> right singular vectors of <img src="https://latex.codecogs.com/png.latex?B"> (GM-selected directions) and <img src="https://latex.codecogs.com/png.latex?V_2"> are the top-<img src="https://latex.codecogs.com/png.latex?k"> PCA directions of FNC. The <strong>subspace overlap</strong> is:</p>
<p><img src="https://latex.codecogs.com/png.latex?%0AO(V_1,%20V_2)%20=%20%5Cfrac%7B1%7D%7Bk%7D%5Csum_%7Bi=1%7D%5E%7Bk%7D%20%5Ccos%5E2%5Ctheta_i%20%5C;%5Cin%20%5B0,%201%5D%0A"></p>
</section>
<section id="the-core-finding-o-0.45-vs.-r2-0.06" class="level2">
<h2 class="anchored" data-anchor-id="the-core-finding-o-0.45-vs.-r2-0.06">The Core Finding: <img src="https://latex.codecogs.com/png.latex?O%20=%200.45"> vs.&nbsp;<img src="https://latex.codecogs.com/png.latex?R%5E2%20=%200.06"></h2>
<p>Across three datasets (schizophrenia cohort <img src="https://latex.codecogs.com/png.latex?N=1%7B,%7D151">; external validation <img src="https://latex.codecogs.com/png.latex?N=102">; UK Biobank <img src="https://latex.codecogs.com/png.latex?N%20%5Capprox%2037%7B,%7D775">):</p>
<table class="caption-top table">
<colgroup>
<col style="width: 25%">
<col style="width: 22%">
<col style="width: 51%">
</colgroup>
<thead>
<tr class="header">
<th>Metric</th>
<th>Value</th>
<th>Interpretation</th>
</tr>
</thead>
<tbody>
<tr class="odd">
<td>Variance explained (<img src="https://latex.codecogs.com/png.latex?R%5E2">)</td>
<td>~0.06</td>
<td>GM weakly predicts FNC amplitudes</td>
</tr>
<tr class="even">
<td>Subspace overlap (<img src="https://latex.codecogs.com/png.latex?O">)</td>
<td>0.447</td>
<td>GM strongly constrains FNC directions</td>
</tr>
<tr class="odd">
<td>Top 3 principal angle cosines</td>
<td>0.97, 0.95, 0.86</td>
<td>Near-perfect alignment in 3 dimensions</td>
</tr>
<tr class="even">
<td>4th principal angle</td>
<td>0.28</td>
<td>Sharp dropoff — concentrated coupling</td>
</tr>
<tr class="odd">
<td>Chance overlap (at <img src="https://latex.codecogs.com/png.latex?k=20">)</td>
<td>~0.015</td>
<td>Observed overlap is 30× above chance</td>
</tr>
</tbody>
</table>
<p>The gap is an order of magnitude. Gray matter does not determine <em>how strongly</em> individuals deviate from mean connectivity — but it strongly constrains <em>in which directions</em> that deviation can occur.</p>
<p><strong>The analogy</strong>: anatomy sets the stage geometry; neural dynamics determine how the actors perform on that stage. The theater constrains the repertoire of possible plays without scripting any particular performance.</p>
</section>
<section id="what-the-coupling-modes-look-like" class="level2">
<h2 class="anchored" data-anchor-id="what-the-coupling-modes-look-like">What the Coupling Modes Look Like</h2>
<div class="quarto-figure quarto-figure-center">
<figure class="figure">
<p><img src="https://yuda-bi.com/files/sfcoupling-fig7.png" class="img-fluid figure-img" style="width:90.0%"></p>
<figcaption><strong>Structure-function coupling modes.</strong> Each row shows one SVD mode: left panel shows GM loadings on a dorsal glass brain (node size = loading magnitude, color = network domain), right panel shows the corresponding FNC loading matrix (domain-sorted). Mode 1 (36.5% of coupled variance) captures a global sensorimotor-cognitive axis; Mode 2 (9.2%) isolates default mode and visual interactions; Mode 3 (7.9%) captures cerebellar-cortical coupling.</figcaption>
</figure>
</div>
<p>The SVD decomposition reveals interpretable structure-function coupling modes:</p>
<ul>
<li><strong>Mode 1</strong> (36.5% of coupled variance): A distributed pattern spanning cognitive control, sensorimotor, and visual domains. This is the dominant axis along which GM morphometry shapes functional organization.</li>
<li><strong>Mode 2</strong> (9.2%): Concentrated in default mode and visual network interactions — reflecting the structural basis of resting-state network architecture.</li>
<li><strong>Mode 3</strong> (7.9%): Captures cerebellar-cortical coupling patterns, highlighting the structural underpinning of cerebro-cerebellar communication.</li>
</ul>
<p>These modes are highly stable across random seeds (correlation <img src="https://latex.codecogs.com/png.latex?r%20%3E%200.9"> for the top 3) and replicate across datasets.</p>
</section>
<section id="linearity-the-coupling-really-is-linear" class="level2">
<h2 class="anchored" data-anchor-id="linearity-the-coupling-really-is-linear">Linearity: The Coupling Really Is Linear</h2>
<p>A critical finding: <strong>nonlinear models provide no reliable improvement</strong> over the linear nuclear norm solution.</p>
<ul>
<li>MLP (multilayer perceptron) gains +0.004 <img src="https://latex.codecogs.com/png.latex?R%5E2"> on discovery data but <strong>reverses to −0.009 on external validation</strong></li>
<li>A nonlinear residual model performs worse than nuclear norm alone on both datasets</li>
<li>When initialized from the linear solution, the MLP’s mixing parameter converges to <img src="https://latex.codecogs.com/png.latex?%5Calpha%20=%200.60">, staying close to the linear regime</li>
</ul>
<p>This linearity is not an assumption — it is an empirical finding validated across datasets. The structure-function relationship, at the population level, is well-captured by a linear, low-rank map.</p>
</section>
<section id="clinical-relevance-the-coupled-subspace-carries-disease-information" class="level2">
<h2 class="anchored" data-anchor-id="clinical-relevance-the-coupled-subspace-carries-disease-information">Clinical Relevance: The Coupled Subspace Carries Disease Information</h2>
<p>Decomposing each subject’s FNC into a <strong>structure-coupled</strong> component (<img src="https://latex.codecogs.com/png.latex?y_%5Cmathrm%7Bcoup%7D%20=%20V_r%20V_r%5E%5Ctop%20y">) and a <strong>structure-uncoupled</strong> component (<img src="https://latex.codecogs.com/png.latex?y_%5Cmathrm%7Buncoup%7D%20=%20(I%20-%20V_r%20V_r%5E%5Ctop)%20y">):</p>
<table class="caption-top table">
<thead>
<tr class="header">
<th>Component</th>
<th>SZ classification AUC</th>
</tr>
</thead>
<tbody>
<tr class="odd">
<td>Coupled FNC (rank 38)</td>
<td><strong>0.795</strong> [0.726, 0.857]</td>
</tr>
<tr class="even">
<td>Full FNC</td>
<td>0.773 [0.712, 0.839]</td>
</tr>
<tr class="odd">
<td>Uncoupled FNC</td>
<td>0.728</td>
</tr>
</tbody>
</table>
<p>The structure-coupled component <em>outperforms full FNC</em> for schizophrenia classification. This means that the functional variation constrained by gray matter morphometry preferentially carries clinically relevant information — anatomical structure does not just constrain function; it constrains the <em>clinically informative</em> part of function.</p>
</section>
<section id="physical-and-mathematical-significance" class="level2">
<h2 class="anchored" data-anchor-id="physical-and-mathematical-significance">Physical and Mathematical Significance</h2>
<section id="why-nuclear-norm" class="level3">
<h3 class="anchored" data-anchor-id="why-nuclear-norm">Why Nuclear Norm?</h3>
<p>The nuclear norm <img src="https://latex.codecogs.com/png.latex?%5C%7CB%5C%7C_*"> is the <img src="https://latex.codecogs.com/png.latex?%5Cell_1"> norm of the singular value vector — it induces sparsity in the spectral domain just as LASSO induces sparsity in the coefficient domain. This is the natural regularizer when the underlying relationship is low-rank: it seeks the simplest (lowest effective rank) linear map consistent with the data.</p>
<p>In the landscape of regularization:</p>
<table class="caption-top table">
<colgroup>
<col style="width: 24%">
<col style="width: 57%">
<col style="width: 18%">
</colgroup>
<thead>
<tr class="header">
<th>Method</th>
<th>What it regularizes</th>
<th>Bias</th>
</tr>
</thead>
<tbody>
<tr class="odd">
<td>Ridge (<img src="https://latex.codecogs.com/png.latex?%5C%7CB%5C%7C_F%5E2">)</td>
<td>Shrinks all singular values equally</td>
<td>Preserves rank, weakens all directions</td>
</tr>
<tr class="even">
<td>PLS</td>
<td>Maximizes covariance in few components</td>
<td>Greedily selects modes, poor generalization</td>
</tr>
<tr class="odd">
<td><strong>Nuclear Norm</strong> (<img src="https://latex.codecogs.com/png.latex?%5C%7CB%5C%7C_*">)</td>
<td>Soft-thresholds singular values</td>
<td>Suppresses weak modes, preserves strong ones</td>
</tr>
</tbody>
</table>
<p>Nuclear Norm achieves the highest external generalization (74% retention) vs.&nbsp;PLS (45%) precisely because it soft-thresholds rather than hard-truncates the spectral structure.</p>
</section>
<section id="the-subspace-overlap-as-an-information-geometric-quantity" class="level3">
<h3 class="anchored" data-anchor-id="the-subspace-overlap-as-an-information-geometric-quantity">The Subspace Overlap as an Information-Geometric Quantity</h3>
<p>The subspace overlap <img src="https://latex.codecogs.com/png.latex?O(V_1,%20V_2)"> is the mean squared cosine of the principal angles — equivalently, it is the normalized Frobenius norm of the product of two orthogonal projections:</p>
<p><img src="https://latex.codecogs.com/png.latex?%0AO%20=%20%5Cfrac%7B1%7D%7Bk%7D%5C%7CP_%7BV_1%7D%20P_%7BV_2%7D%5C%7C_F%5E2%0A"></p>
<p>This has a natural interpretation in information geometry: it measures the fraction of “geometric information” shared between two low-dimensional representations of the same high-dimensional space. When <img src="https://latex.codecogs.com/png.latex?O%20%5Cgg%20R%5E2">, the two modalities share geometric structure (directional alignment) without sharing amplitude information — the hallmark of a <em>soft constraint</em> rather than a deterministic prediction.</p>
</section>
</section>
<section id="looking-ahead-the-multimodal-decomposition-program" class="level2">
<h2 class="anchored" data-anchor-id="looking-ahead-the-multimodal-decomposition-program">Looking Ahead: The Multimodal Decomposition Program</h2>
<p>This paper is the first in a planned three-paper series:</p>
<p><strong>Paper 0</strong> (this work): Establishes that GM constrains function through a shared low-rank subspace, not pointwise prediction. Introduces nuclear norm regularization and the <img src="https://latex.codecogs.com/png.latex?O%20%5Cgg%20R%5E2"> diagnostic.</p>
<p><strong>Paper 1</strong> (in preparation): Asks whether gray matter and white matter constrain the <em>same</em> or <em>different</em> functional subspaces. Uses two-stage subspace decomposition to partition functional variance into four components:</p>
<p><img src="https://latex.codecogs.com/png.latex?%0Af_i%20=%20f_i%5E%7B(g%20%5Ccap%20w)%7D%20+%20f_i%5E%7B(g%20%5Csetminus%20w)%7D%20+%20f_i%5E%7B(w%20%5Csetminus%20g)%7D%20+%20f_i%5E%7B(%5Cmathrm%7Bres%7D)%7D%0A"></p>
<p>Key question: is the GM–WM functional overlap redundant or complementary? Data: UK Biobank (<img src="https://latex.codecogs.com/png.latex?N%20%5Capprox%2030%7B,%7D000">; sMRI + dMRI + resting fMRI).</p>
<p><strong>Paper 2</strong> (planned): Unified probabilistic framework with joint latent variables, non-Gaussian priors for identifiability, and posterior uncertainty on the variance decomposition. Connects to ICA/IVA frameworks (Adalı) and Bayesian ARD for automatic dimensionality selection.</p>
<hr>
</section>
<section id="paper" class="level2">
<h2 class="anchored" data-anchor-id="paper">Paper</h2>
<p><strong>Yuda Bi</strong>, Vince D. Calhoun. <em>Gray Matter Morphometry Reveals a Soft Low-Rank Structure-Function Subspace.</em> In preparation for NeuroImage.</p>


</section>

 ]]></description>
  <category>neuroscience</category>
  <category>structure-function coupling</category>
  <category>spectral methods</category>
  <category>linear algebra</category>
  <guid>https://yuda-bi.com/notes/2026-04-sfcoupling/</guid>
  <pubDate>Tue, 07 Apr 2026 00:00:00 GMT</pubDate>
  <media:content url="https://yuda-bi.com/notes/2026-04-sfcoupling/featured.png" medium="image" type="image/png" height="45" width="144"/>
</item>
</channel>
</rss>
