<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>Theory on Hunter Heidenreich | ML Research Scientist</title><link>https://hunterheidenreich.com/tags/theory/</link><description>Recent content in Theory on Hunter Heidenreich | ML Research Scientist</description><image><title>Hunter Heidenreich | ML Research Scientist</title><url>https://hunterheidenreich.com/img/avatar.webp</url><link>https://hunterheidenreich.com/img/avatar.webp</link></image><generator>Hugo -- 0.147.7</generator><language>en-US</language><copyright>2026 Hunter Heidenreich</copyright><lastBuildDate>Sat, 11 Apr 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://hunterheidenreich.com/tags/theory/index.xml" rel="self" type="application/rss+xml"/><item><title>Defining Disentangled Representations via Group Theory</title><link>https://hunterheidenreich.com/notes/machine-learning/geometric-deep-learning/defining-disentangled-representations/</link><pubDate>Sat, 11 Apr 2026 00:00:00 +0000</pubDate><guid>https://hunterheidenreich.com/notes/machine-learning/geometric-deep-learning/defining-disentangled-representations/</guid><description>First formal definition of disentangled representations using group theory, connecting symmetry transformations to vector space decompositions.</description><content:encoded><![CDATA[<h2 id="a-theory-paper-grounding-disentanglement-in-symmetry">A Theory Paper Grounding Disentanglement in Symmetry</h2>
<p>This is a <strong>Theory</strong> paper that provides the first formal mathematical definition of disentangled representations. Rather than proposing a new learning algorithm or evaluating existing methods, the paper uses group theory and representation theory to define precisely what it means for a representation to be disentangled. The authors argue that the relevant structure of the world is captured by symmetry transformations, and that a disentangled representation must decompose into independent subspaces aligned with the decomposition of the corresponding symmetry group.</p>
<h2 id="why-disentangling-lacks-a-formal-foundation">Why Disentangling Lacks a Formal Foundation</h2>
<p>Disentangled representation learning aims to learn representations where distinct factors of variation in the data are separated into independent components. This idea has driven significant research, particularly through models like $\beta$-VAE and InfoGAN. Despite this progress, the field has lacked agreement on several fundamental questions: what constitutes the &ldquo;data generative factors,&rdquo; whether each factor should correspond to a single latent dimension or multiple dimensions, and whether a disentangled representation should have a unique axis alignment.</p>
<p>Without a formal definition, evaluating disentanglement methods remains subjective, relying on human intuition or metrics that encode different (sometimes contradictory) assumptions. For example, some metrics penalize multi-dimensional subspaces while others allow them. The lack of formal grounding also means there is no principled way to determine whether certain factors of variation (such as 3D rotations) can even be disentangled in principle.</p>
<p>The authors draw inspiration from physics, where symmetry transformations have been central to understanding world structure since <a href="https://en.wikipedia.org/wiki/Noether%27s_theorem">Noether&rsquo;s theorem</a> connected conservation laws to continuous symmetries. Gell-Mann&rsquo;s prediction of the $\Omega^{-}$ particle from symmetry-based classification of hadrons, and the unification of electricity and magnetism through shared symmetry transformations, illustrate the power of the symmetry perspective for generalization to new domains.</p>
<h2 id="symmetry-groups-as-the-foundation-for-disentanglement">Symmetry Groups as the Foundation for Disentanglement</h2>
<p>The core insight is that the &ldquo;data generative factors&rdquo; previously used to discuss disentanglement should be replaced by symmetry transformations of the world. The paper defines a disentangled representation through three key concepts.</p>
<h3 id="disentangled-group-action">Disentangled Group Action</h3>
<p>Given a group $G$ that decomposes as a <a href="https://en.wikipedia.org/wiki/Direct_product_of_groups">direct product</a> $G = G_1 \times G_2 \times \ldots \times G_n$, an action of $G$ on a set $X$ is <strong>disentangled</strong> if there exists a decomposition $X = X_1 \times X_2 \times \ldots \times X_n$ such that each subgroup $G_i$ acts only on $X_i$ and leaves all other components fixed:</p>
<p>$$(g_1, g_2) \cdot (v_1, v_2) = (g_1 \cdot_1 v_1, g_2 \cdot_2 v_2)$$</p>
<h3 id="disentangled-representation">Disentangled Representation</h3>
<p>Let $W$ be the set of world states with symmetry group $G$ acting on it. A generative process $b: W \to O$ produces observations, and an inference process $h: O \to Z$ produces representations. The composition $f = h \circ b$ maps world states to representations. The representation is <strong>disentangled</strong> if:</p>
<ol>
<li>There exists an action $\cdot: G \times Z \to Z$</li>
<li>The map $f: W \to Z$ is <strong><a href="https://en.wikipedia.org/wiki/Equivariant_map">equivariant</a></strong>: $g \cdot f(w) = f(g \cdot w)$ for all $g \in G, w \in W$</li>
<li>There exists a decomposition $Z = Z_1 \oplus Z_2 \oplus \ldots \oplus Z_n$ such that each $Z_i$ is affected only by $G_i$ and fixed by all other subgroups</li>
</ol>
<p>The equivariance condition ensures that the symmetry structure of the world is faithfully reflected in the representation space.</p>
<h3 id="linear-disentangled-representation">Linear Disentangled Representation</h3>
<p>When the group action on $Z$ is additionally constrained to be linear, the representation becomes a <strong>linear disentangled representation</strong>. This leverages <a href="https://en.wikipedia.org/wiki/Group_representation">group representation theory</a>, where the action is described by a homomorphism $\rho: G \to GL(Z)$. The representation is linearly disentangled if it decomposes as a direct sum $\rho = \rho_1 \oplus \rho_2 \oplus \ldots \oplus \rho_n$, where each $\rho_i$ acts only on $Z_i$. In matrix terms, this means $\rho(g)$ takes a block-diagonal form.</p>
<p>For the irreducible representations of a direct product group $G = G_1 \times G_2$, disentanglement requires that each irreducible component $\rho_1 \otimes \rho_2$ has at most one non-trivial factor. This prevents any subspace from being jointly affected by multiple subgroups.</p>
<h2 id="grid-world-example-and-the-so3-counterexample">Grid World Example and the SO(3) Counterexample</h2>
<p>Since this is a theory paper, the &ldquo;experiments&rdquo; consist of worked examples that illustrate the definition.</p>
<h3 id="grid-world-verification">Grid World Verification</h3>
<p>The authors consider a grid world where an object can translate horizontally, vertically, and change color, with wraparound boundaries. The symmetry group decomposes as $G = G_x \times G_y \times G_c$, where each subgroup is isomorphic to the <a href="https://en.wikipedia.org/wiki/Cyclic_group">cyclic group</a> $C_N$.</p>
<p>A CCI-VAE model trained on observations from this world learns a representation that approximately satisfies the equivariance condition $f(x, y, c) \approx (\lambda_x x, \lambda_y y, \lambda_c c)$, where each subgroup acts independently on its corresponding subspace. The group structure (commutativity of actions) is approximately preserved, though the learned representation uses translation rather than linear action, and the cyclic structure is lost.</p>
<p>For a linear disentangled representation, the map $f(x, y, c) = (e^{2\pi i x / N}, e^{2\pi i y / N}, e^{2\pi i c / N})$ over $\mathbb{C}^3$ provides an exact solution. The generator of each subgroup acts as multiplication by $e^{2\pi i / N}$ on its corresponding coordinate, yielding a truly linear and disentangled action. Equivalently, viewing $\rho$ as a representation over $\mathbb{R}^6$ (since $\mathbb{C}^3 \cong \mathbb{R}^6$), the group action is expressed using block-diagonal matrices of $2 \times 2$ rotation matrices, and each invariant subspace becomes two-dimensional.</p>
<h3 id="3d-rotations-cannot-be-disentangled">3D Rotations Cannot Be Disentangled</h3>
<p>The group of 3D rotations <a href="https://en.wikipedia.org/wiki/3D_rotation_group">$SO(3)$</a> has subgroups for rotations about the $x$, $y$, and $z$ axes. Intuitively, one might expect to disentangle these three rotation axes. However, rotations about different axes do not commute (rotating $90°$ about $x$ then $y$ gives a different result from $y$ then $x$), so $SO(3)$ cannot be written as a direct product of these subgroups. The definition correctly rules out disentangling along these lines.</p>
<p>Rotations can still be disentangled from other independent symmetries. For an object that can rotate and change color, the relevant group $G = SO(3) \times G_c$ is a valid direct product, so rotation and color form two disentangled subspaces (even though the rotation subspace is itself multi-dimensional and internally entangled).</p>
<h2 id="resolving-disagreements-and-defining-the-path-forward">Resolving Disagreements and Defining the Path Forward</h2>
<h3 id="backward-compatibility-with-existing-intuitions">Backward Compatibility with Existing Intuitions</h3>
<p>The paper evaluates its definition against three established dimensions of disentanglement:</p>
<p><strong>Modularity</strong> (each latent dimension encodes at most one factor): Satisfied by the new definition, with &ldquo;data generative factors&rdquo; replaced by &ldquo;disentangled actions of the symmetry group.&rdquo; The $SO(3)$ case shows where the new definition disagrees with naive intuition, correctly identifying that non-commuting factors cannot be disentangled.</p>
<p><strong>Compactness</strong> (each factor encoded by a single dimension): The new definition allows multi-dimensional subspaces, siding with approaches that permit distributed representations of individual factors. The dimensionality of each subspace is determined by the structure of the corresponding group representation.</p>
<p><strong>Explicitness</strong> (factors linearly decodable): The general definition does not require linearity. Linear disentangled representations are a strictly stronger condition, and the paper provides a separate formal definition for this case.</p>
<h3 id="key-consequences">Key Consequences</h3>
<p>The definition is relative to a particular decomposition of the symmetry group into subgroups. This has two implications. First, the same group may admit multiple decompositions, and different decompositions yield different disentangled representations (potentially useful for different downstream tasks). Second, identifying the &ldquo;natural&rdquo; decomposition is a separate problem that the authors leave to future work, suggesting that active perception and causal interventions may play a role.</p>
<p>The paper connects to Locatello et al. (2018), who proved that unsupervised learning of disentangled representations is impossible without inductive biases. The symmetry-based framework suggests that such biases could come from an agent&rsquo;s ability to interact with the world and discover which aspects remain invariant under various transformations.</p>
<h3 id="limitations">Limitations</h3>
<p>The paper explicitly focuses on defining disentanglement rather than solving the learning problem. It assumes that the symmetry group decomposes as a direct product of subgroups and that a useful decomposition is known. The authors acknowledge that relaxing these assumptions (e.g., discovering useful decompositions automatically) is important future work. The worked examples use toy environments, and bridging the gap to realistic data remains an open challenge.</p>
<hr>
<h2 id="reproducibility-details">Reproducibility Details</h2>
<h3 id="data">Data</h3>
<p>This is a purely theoretical paper. The only empirical element is a qualitative demonstration using a CCI-VAE model on a grid world environment, where an object translates on a grid with wraparound and changes color through discrete steps on a circular hue axis.</p>
<h3 id="algorithms">Algorithms</h3>
<p>No new algorithms are proposed. The CCI-VAE model from Burgess et al. (2018) is used for the grid world demonstration. The paper&rsquo;s contribution is a set of formal definitions, not an algorithmic procedure.</p>
<h3 id="evaluation">Evaluation</h3>
<p>No quantitative evaluation is performed. The paper discusses how existing disentanglement metrics relate to the proposed definition, noting that they each capture different subsets of the three dimensions (modularity, compactness, explicitness) and that the formal definition provides a principled way to evaluate their relative merits.</p>
<h3 id="reproducibility-status-closed">Reproducibility Status: Closed</h3>
<p>This is a theory paper whose primary contribution is a set of formal definitions. The theoretical content (definitions, proofs, worked examples) is self-contained in the paper. No code, data, or models are released. The CCI-VAE demonstration uses a model from Burgess et al. (2018), but no implementation or training details specific to the grid world experiment are provided.</p>
<hr>
<h2 id="paper-information">Paper Information</h2>
<p><strong>Citation</strong>: Higgins, I., Amos, D., Pfau, D., Racanière, S., Matthey, L., Rezende, D., &amp; Lerchner, A. (2018). Towards a Definition of Disentangled Representations. <em>arXiv preprint arXiv:1812.02230</em>.</p>
<div class="highlight"><pre tabindex="0" style="color:#f8f8f2;background-color:#272822;-moz-tab-size:4;-o-tab-size:4;tab-size:4;"><code class="language-bibtex" data-lang="bibtex"><span style="display:flex;"><span><span style="color:#a6e22e">@misc</span>{higgins2018towards,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">title</span>=<span style="color:#e6db74">{Towards a Definition of Disentangled Representations}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">author</span>=<span style="color:#e6db74">{Higgins, Irina and Amos, David and Pfau, David and Racani\`{e}re, S\&#39;{e}bastien and Matthey, Lo\&#34;{i}c and Rezende, Danilo and Lerchner, Alexander}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">year</span>=<span style="color:#e6db74">{2018}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">eprint</span>=<span style="color:#e6db74">{1812.02230}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">archiveprefix</span>=<span style="color:#e6db74">{arXiv}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">primaryclass</span>=<span style="color:#e6db74">{cs.LG}</span>
</span></span><span style="display:flex;"><span>}
</span></span></code></pre></div>]]></content:encoded></item><item><title>Atom-Density Representations for Machine Learning</title><link>https://hunterheidenreich.com/notes/chemistry/molecular-simulation/ml-potentials/atom-density-representations-ml/</link><pubDate>Sat, 11 Apr 2026 00:00:00 +0000</pubDate><guid>https://hunterheidenreich.com/notes/chemistry/molecular-simulation/ml-potentials/atom-density-representations-ml/</guid><description>A unified bra-ket framework connecting SOAP, Behler-Parrinello, and other atom-density representations for ML on molecules and materials.</description><content:encoded><![CDATA[<h2 id="a-unified-theory-of-atom-density-representations">A Unified Theory of Atom-Density Representations</h2>
<p>This is a <strong>Theory</strong> paper that provides a formal, basis-independent framework for constructing structural representations of atomic systems for machine learning. Rather than proposing a new representation, Willatt, Musil, and Ceriotti show that many popular approaches (SOAP power spectra, Behler-Parrinello symmetry functions, $n$-body kernels, and tensorial SOAP) are special cases of a single abstract construction based on smoothed atom densities and <a href="https://en.wikipedia.org/wiki/Haar_measure">Haar integration</a> over symmetry groups.</p>
<h2 id="the-challenge-of-representing-atomic-structures">The Challenge of Representing Atomic Structures</h2>
<p>Machine learning models for predicting molecular and materials properties require input representations that are (1) complete enough to distinguish structurally distinct configurations and (2) invariant to physical symmetries (translations, rotations, and permutations of identical atoms). This has led to a large and growing set of competing approaches: <a href="/posts/molecular-descriptor-coulomb-matrix/">Coulomb matrices</a>, symmetry functions, <a href="https://en.wikipedia.org/wiki/Radial_distribution_function">radial distribution functions</a>, wavelets, invariant polynomials, and many more.</p>
<p>The proliferation of representations makes it difficult to compare them on equal footing or to identify which design choices are fundamental and which are incidental. Internal-coordinate approaches (e.g., Coulomb matrices) are automatically translation- and rotation-invariant but require additional symmetrization over permutations, which can introduce derivative discontinuities when done via sorting. Density-based approaches such as radial distribution functions and SOAP avoid these discontinuities by working with smooth density fields, but their theoretical connections to one another have not been made explicit.</p>
<h2 id="dirac-notation-for-atomic-environments">Dirac Notation for Atomic Environments</h2>
<p>The core innovation is to describe an atomic configuration $\mathcal{A}$ as a ket $|\mathcal{A}\rangle$ in a Hilbert space, formed by placing smooth functions $g(\mathbf{r})$ (typically Gaussians) on each atom and decorating them with orthonormal element kets $|\alpha\rangle$:</p>
<p>$$
\langle \mathbf{r} | \mathcal{A} \rangle = \sum_{i} g(\mathbf{r} - \mathbf{r}_{i}) | \alpha_{i} \rangle
$$</p>
<p>This ket is basis-independent, which is the reason for adopting the Dirac notation. The same abstract object can be projected onto position space, reciprocal space, or a basis of radial functions and spherical harmonics, yielding different concrete representations that all encode the same structural information.</p>
<h3 id="symmetrization-via-haar-integration">Symmetrization via Haar Integration</h3>
<p>To impose translational invariance, the ket is averaged over the translation group. Averaging the raw density $|\mathcal{A}\rangle$ directly (first order, $\nu = 1$) discards all geometric information and retains only atom counts per element. The solution is to first take tensor products and then average:</p>
<p>$$
\left| \mathcal{A}^{(\nu)} \right\rangle_{\hat{t}} = \int \mathrm{d}\hat{t} \underbrace{\hat{t}|\mathcal{A}\rangle \otimes \hat{t}|\mathcal{A}\rangle \cdots \hat{t}|\mathcal{A}\rangle}_{\nu}
$$</p>
<p>For $\nu = 2$, this yields a translationally-invariant ket that encodes pairwise distance information between atoms, and naturally decomposes into atom-centered contributions:</p>
<p>$$
\left| \mathcal{A}^{(2)} \right\rangle_{\hat{t}} = \sum_{j} |\alpha_{j}\rangle |\mathcal{X}_{j}\rangle
$$</p>
<p>where $|\mathcal{X}_{j}\rangle$ is the environment ket centered on atom $j$, defined with a smooth cutoff function $f_{c}(r_{ij})$ that restricts each environment to a spherical neighborhood (justified by the nearsightedness principle of electronic matter). This decomposition is what justifies the widely used additive kernel between structures (a sum of kernels between environments).</p>
<h3 id="rotational-invariance-and-body-order-correlations">Rotational Invariance and Body-Order Correlations</h3>
<p>The same Haar integration procedure over the $SO(3)$ rotation group produces rotationally invariant representations:</p>
<p>$$
\left| \mathcal{X}^{(\nu)} \right\rangle_{\hat{R}} = \int \mathrm{d}\hat{R} \prod_{\aleph}^{\nu} \otimes \hat{R}\hat{U}_{\aleph}|\mathcal{X}_{j}^{\aleph}\rangle
$$</p>
<p>The order $\nu$ of the tensor product before symmetrization determines the body-order of correlations captured: $\nu$ corresponds to $(\nu + 1)$-body correlations. The $\nu = 1$ invariant ket retains only radial (distance) information (two-body). The $\nu = 2$ ket encodes three-body correlations (two distances and an angle), and is argued to be sufficient for unique reconstruction of a configuration (up to inversion symmetry), based on extensive numerical experiments. Using nonlinear kernels (tensor products of the symmetrized ket, parameterized by $\zeta$) allows the model to incorporate higher body-order correlations beyond those explicitly in the feature vector.</p>
<h2 id="recovering-soap-symmetry-functions-and-tensorial-extensions">Recovering SOAP, Symmetry Functions, and Tensorial Extensions</h2>
<p>By projecting the abstract invariant kets onto specific basis sets, the authors recover several well-known frameworks as special cases.</p>
<h3 id="behler-parrinello-symmetry-functions">Behler-Parrinello Symmetry Functions</h3>
<p>In the $\delta$-function limit of the atomic density, the $\nu = 1$ and $\nu = 2$ invariant kets in real space directly correspond to the 2-body and 3-body correlation functions. Behler-Parrinello symmetry functions are projections of these correlation functions onto suitable test functions $G$:</p>
<p>$$
\langle \alpha \beta G_{2} | \mathcal{X}_{j} \rangle = \langle \alpha | \alpha_{j} \rangle \int \mathrm{d}r, G_{2}(r), r \left\langle \beta r | \mathcal{X}_{j}^{(1)} \right\rangle_{\hat{R},, h \to \delta}
$$</p>
<p>where the $h \to \delta$ subscript indicates the Dirac delta limit of the atomic density.</p>
<h3 id="soap-power-spectrum">SOAP Power Spectrum</h3>
<p>Expanding the environmental ket in a basis of radial functions $R_{n}(r)$ and <a href="https://en.wikipedia.org/wiki/Spherical_harmonics">spherical harmonics</a> $Y_{m}^{l}(\hat{\mathbf{r}})$, the $\nu = 2$ invariant ket is the SOAP power spectrum:</p>
<p>$$
\left\langle \alpha n \alpha&rsquo; n&rsquo; l \left| \mathcal{X}_{j}^{(2)} \right\rangle_{\hat{R}} \propto \frac{1}{\sqrt{2l+1}} \sum_{m} \langle \alpha n l m | \mathcal{X}_{j} \rangle^{\star} \langle \alpha&rsquo; n&rsquo; l m | \mathcal{X}_{j} \rangle \right.
$$</p>
<p>This identity shows that the SOAP kernel, which can be expressed as a scalar product between truncated power spectrum vectors, is a natural consequence of the inner product between invariant kets. The $\nu = 3$ case yields the <a href="https://en.wikipedia.org/wiki/Bispectrum">bispectrum</a>, used as a four-body feature vector in both SOAP and Spectral Neighbor Analysis Potentials (SNAP), where its high resolution enables accurate interatomic potentials through linear regression:</p>
<p>$$
\langle \alpha_{1} n_{1} l_{1}, \alpha_{2} n_{2} l_{2}, \alpha n l | \mathcal{X}_{j}^{(3)} \rangle_{\hat{R}} \propto \frac{1}{\sqrt{2l+1}} \sum_{m, m_{1}, m_{2}} \langle \mathcal{X}_{j} | \alpha n l m \rangle \langle \alpha_{1} n_{1} l_{1} m_{1} | \mathcal{X}_{j} \rangle \langle \alpha_{2} n_{2} l_{2} m_{2} | \mathcal{X}_{j} \rangle \langle l_{1} m_{1} l_{2} m_{2} | l m \rangle
$$</p>
<p>where $\langle l_{1} m_{1} l_{2} m_{2} | l m \rangle$ is a <a href="https://en.wikipedia.org/wiki/Clebsch%E2%80%93Gordan_coefficients">Clebsch-Gordan coefficient</a>.</p>
<h3 id="tensorial-soap-lambda-soap">Tensorial SOAP ($\lambda$-SOAP)</h3>
<p>The tensorial extension of SOAP incorporates an angular momentum ket $|\lambda \mu\rangle$ into the tensor product before symmetrization:</p>
<p>$$
\left| \mathcal{X}^{(\nu)} \lambda \mu \right\rangle_{\hat{R}} = \int \mathrm{d}\hat{R}, \hat{R}|\lambda \mu\rangle \prod_{\aleph=1}^{\nu} \otimes \hat{R}|\mathcal{X}_{j}\rangle
$$</p>
<p>This construction is rotationally invariant in the full product space but covariant in the subspace of atomic environments, enabling models for tensorial properties (e.g., polarizability tensors, chemical shielding).</p>
<h3 id="distributions-vs-sorted-vectors">Distributions vs. Sorted Vectors</h3>
<p>The paper also connects density-based and sorted-vector approaches. Given a set of structural descriptors ${a_{i}}$, the sorted vector is equivalent to the inverse cumulative distribution function of the histogram of values. The Euclidean distance between sorted vectors is the $\mathcal{L}^{2}$ norm of the difference between the inverse CDFs, and the $\mathcal{L}^{1}$ norm corresponds to the <a href="https://en.wikipedia.org/wiki/Earth_mover%27s_distance">earth mover&rsquo;s distance</a>. This highlights that different symmetrization strategies encode essentially the same structural information.</p>
<h2 id="generalized-operators-for-tuning-representations">Generalized Operators for Tuning Representations</h2>
<p>The framework becomes especially powerful through the introduction of a linear Hermitian operator $\hat{U}$ that transforms the density ket before symmetrization. This operator must commute with rotations:</p>
<p>$$
\langle \alpha n l m | \hat{U} | \alpha&rsquo; n&rsquo; l&rsquo; m&rsquo; \rangle = \delta_{ll&rsquo;} \delta_{mm&rsquo;} \langle \alpha n l | \hat{U} | \alpha&rsquo; n&rsquo; l&rsquo; \rangle
$$</p>
<p>Several practical modifications to standard representations can be understood as choices of $\hat{U}$:</p>
<h3 id="dimensionality-reduction">Dimensionality Reduction</h3>
<p>A low-rank expansion of $\hat{U}$ via PCA on the spherical-harmonic covariance matrix of environments identifies linearly independent components, enabling compression of the feature vector. For a given $l$, the covariance matrix between spherical expansion coefficients is:</p>
<p>$$
C_{\alpha n \alpha&rsquo; n&rsquo;}^{(l)} = \frac{1}{N} \sum_{j} \sum_{m} \langle \alpha n l m | \mathcal{X}_{j} \rangle^{\star} \langle \mathcal{X}_{j} | \alpha&rsquo; n&rsquo; l m \rangle = \frac{\sqrt{2l+1}}{N} \sum_{j} \left\langle \alpha n, \alpha&rsquo; n&rsquo; l \left| \mathcal{X}_{j}^{(2)} \right\rangle_{\hat{R}} \right.
$$</p>
<p>The eigenvectors of $\mathbf{C}^{(l)}$ provide the mixing coefficients for a compressed representation, retaining only components with significant eigenvalues.</p>
<h3 id="radial-scaling">Radial Scaling</h3>
<p>In systems with relatively uniform atom density, the overlap kernel is dominated by the region farthest from the center. A radial scaling operator $u(r)$ (diagonal in position space) downweights distant contributions:</p>
<p>$$
\langle \alpha \mathbf{r} | \hat{U} | \mathcal{X}_{j} \rangle = u(r), \psi_{\mathcal{X}_{j}}^{\alpha}(\mathbf{r})
$$</p>
<p>This recovers the multi-scale kernels that are known to improve predictions in practice, and connects to the two-body features of Faber et al.</p>
<h3 id="alchemical-kernels">Alchemical Kernels</h3>
<p>An operator that acts only in chemical-element space introduces correlations between different elements. The &ldquo;alchemical&rdquo; projection:</p>
<p>$$
\langle J \mathbf{r} | \mathcal{X}_{j} \rangle = \sum_{\alpha} u_{J\alpha}, \psi_{\mathcal{X}_{j}}^{\alpha}(\mathbf{r})
$$</p>
<p>reduces the dimensionality from $O(n_{\mathrm{sp}}^{2})$ to $O(d_{J}^{2})$ and has been shown to produce a low-dimensional representation of elemental space that shares similarities with periodic-table groupings.</p>
<h3 id="non-factorizable-operators">Non-Factorizable Operators</h3>
<p>For more complex modifications (e.g., distance- and angle-dependent scaling of three-body correlations), the operator must act on the full product space rather than factoring into independent components. The authors show that the three-body scaling function of Faber et al. corresponds to a diagonal non-factorizable operator in the real-space representation.</p>
<h2 id="implications-and-future-directions">Implications and Future Directions</h2>
<p>The main conclusions are:</p>
<ol>
<li>
<p><strong>Unification</strong>: SOAP, Behler-Parrinello symmetry functions, $\lambda$-SOAP, bispectrum descriptors, and sorted-vector approaches all emerge from the same abstract construction, differing only in the choice of basis set, body order $\nu$, and kernel power $\zeta$.</p>
</li>
<li>
<p><strong>Systematic improvability</strong>: The $\hat{U}$ operator framework provides a principled way to tune representations, from simple radial scaling to full alchemical and non-factorizable couplings, with clear connections to existing heuristic modifications.</p>
</li>
<li>
<p><strong>Completeness hierarchy</strong>: The body-order parameter $\nu$ and kernel power $\zeta$ together control the trade-off between completeness and computational cost. The $\nu = 2$ (three-body) representation appears to be sufficient for unique structural identification, while higher orders can be recovered through nonlinear kernels.</p>
</li>
</ol>
<p><strong>Limitations</strong>: The paper is primarily theoretical and does not include extensive numerical benchmarks comparing the different instantiations of the framework. Optimization of the $\hat{U}$ operator (especially in its general form) carries a risk of overfitting that the authors acknowledge but do not resolve. The connection to neural network-based representations (message-passing networks, equivariant architectures) is not explored.</p>
<hr>
<h2 id="reproducibility-details">Reproducibility Details</h2>
<h3 id="data">Data</h3>
<p>This is a theoretical paper that does not introduce new benchmark results. No training or test datasets are used. The ethanol molecule in Figure 2 serves as a visualization example for three-body correlation functions.</p>
<h3 id="algorithms">Algorithms</h3>
<p>The paper defines abstract constructions (Haar integration, tensor products, operator transformations) and shows how they reduce to concrete algorithms:</p>
<ul>
<li><strong>SOAP power spectrum</strong>: Expand atom density in radial functions and spherical harmonics, compute $\nu = 2$ invariant ket (Eq. 33).</li>
<li><strong>Alchemical projection</strong>: Contract element channels via learned or PCA-derived mixing coefficients (Eq. 53-55).</li>
<li><strong>Dimensionality reduction</strong>: PCA on the covariance matrix $C_{\alpha n \alpha&rsquo; n&rsquo;}^{(l)}$ of spherical expansion coefficients (Eq. 45).</li>
</ul>
<h3 id="models">Models</h3>
<p>No trained models are presented. The framework applies to kernel ridge regression / Gaussian process regression models that use these representations as inputs.</p>
<h3 id="evaluation">Evaluation</h3>
<p>No quantitative benchmarks are reported. The contribution is the theoretical framework itself, connecting and generalizing existing representations.</p>
<h3 id="hardware">Hardware</h3>
<p>Not applicable (theoretical work).</p>
<hr>
<h2 id="paper-information">Paper Information</h2>
<p><strong>Citation</strong>: Willatt, M. J., Musil, F., &amp; Ceriotti, M. (2019). Atom-density representations for machine learning. <em>The Journal of Chemical Physics</em>, 150(15), 154110. <a href="https://doi.org/10.1063/1.5090481">https://doi.org/10.1063/1.5090481</a></p>
<p><strong>Publication</strong>: The Journal of Chemical Physics, 2019</p>
<div class="highlight"><pre tabindex="0" style="color:#f8f8f2;background-color:#272822;-moz-tab-size:4;-o-tab-size:4;tab-size:4;"><code class="language-bibtex" data-lang="bibtex"><span style="display:flex;"><span><span style="color:#a6e22e">@article</span>{willatt2019atom,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">title</span>=<span style="color:#e6db74">{Atom-density representations for machine learning}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">author</span>=<span style="color:#e6db74">{Willatt, Michael J. and Musil, F{\&#39;e}lix and Ceriotti, Michele}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">journal</span>=<span style="color:#e6db74">{The Journal of Chemical Physics}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">volume</span>=<span style="color:#e6db74">{150}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">number</span>=<span style="color:#e6db74">{15}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">pages</span>=<span style="color:#e6db74">{154110}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">year</span>=<span style="color:#e6db74">{2019}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">doi</span>=<span style="color:#e6db74">{10.1063/1.5090481}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">eprint</span>=<span style="color:#e6db74">{1807.00408}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">archiveprefix</span>=<span style="color:#e6db74">{arXiv}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">primaryclass</span>=<span style="color:#e6db74">{physics.chem-ph}</span>
</span></span><span style="display:flex;"><span>}
</span></span></code></pre></div>]]></content:encoded></item></channel></rss>