A Unified Tensorial Framework for Selection Theory: Generalizing the Price Equation with Applications to Natural, Sexual, and Multistage Selection Dynamics
1. Introduction
In his posthumously published manuscript, George R. Price [1] argued for a general mathematical theory of selection, analogous to Shannon's theory of communication [4]. Price observed that selection operates not only in genetics but in chemistry, linguistics, psychology, economics, and beyond, yet no unified formalism existed to describe it. He proposed a general selection model (Fig. 4 of [1]) in which a pre-selection set \(P\) of entities with amounts \(w_i\) and properties \(x_i\) is transformed into a corresponding set \(P'\) with amounts \(w_i'\) and properties \(x_i'\), such that the \(w_i'\) are non-randomly related to the \(x_i\). From this, Price derived the now-famous Price equation [2,3]:
where the first term represents selection (covariance between fitness \(w\) and trait \(x\)) and the second term represents transmission or property change.
Despite its generality, the classical Price equation remains primarily algebraic and scalar in form. It does not explicitly handle multiple traits, nonlinear interactions, or the composition of different selection mechanisms—such as natural selection (differential survival/reproduction) and sexual selection (differential mating success). Price himself noted that "selection theory is a theory waiting to be born" and called for a more abstract, mathematical treatment akin to Shannon's information theory.
In this work, we answer Price's call by reformulating selection theory in the language of tensors—multilinear maps between vector spaces that are invariant under coordinate changes. Tensors provide a natural framework for describing how traits, fitness, and selective pressures interact in high-dimensional phenotype spaces. We define a selection tensor \(\mathcal{S}\) that generalizes the covariance term of the Price equation to multiple traits and selective forces, and a transmission tensor \(\mathcal{T}\) that generalizes the expectation term. Together, they form a tensorial Price equation:
where \(\bar{\mathbf{x}}\) is a trait vector and \(\cdot\) denotes appropriate tensor contraction.
This formulation allows us to:
- Unify different selection modes (natural, sexual, subset, cultural) as special cases of tensor operators.
- Model multiple, possibly interacting traits and selective pressures.
- Incorporate nonlinear and frequency-dependent selection via curvature terms in the trait manifold.
- Compose selection processes via tensor products, reflecting Price's insight that selection can be multistage (Fig. 6a of [1]).
- Clarify the relationship between selection and entropy via connections to information geometry.
- Handle both discrete generational changes (exactly) and continuous-time limits (approximately).
We begin by reviewing Price's general selection model and its limitations (Section 2), then develop the tensorial framework (Section 3), derive special cases for different selection modes (Section 4), develop dynamical systems for multistage selection (Section 5), integrate Kalman filtering for uncertainty quantification (Section 6), present numerical examples (Section 7), and discuss implications for a general "mathematical theory of selection" (Section 8).
2. Price's General Selection Model and the Need for Tensors
Price's model [1, Fig. 4] represents selection as a mapping from a pre-selection population \(P = \{(w_i, x_i)\}\) to a post-selection population \(P' = \{(w_i', x_i')\}\), with a one-to-one correspondence between elements. The selection coefficient is \(s_i = w_i' / w_i\), and selection is defined as a non-random relationship between \(s_i\) and \(x_i\). The model accommodates:
- Subset selection (choosing apples by quality) via empty packages (\(w_i' = 0\) for unselected).
- Darwinian natural selection (differential reproduction) via \(w_i' > w_i\) and \(x_i' \neq x_i\).
- Multi-stage selection (Fig. 6a of [1]) with alternating property change and selection steps.
Price noted that the covariance form \(\Delta \bar{x} = \operatorname{Cov}(w,x)/\bar{w}\) emerges naturally from this model, but he also identified complexities not captured by the scalar equation:
- Multiple traits: Real selection acts on many traits simultaneously, possibly with correlations.
- Multiple selection criteria: Sexual selection, natural selection, and artificial selection may act concurrently.
- Nonlinear interactions: Fitness may depend nonlinearly on traits or trait combinations.
- Property change during transmission: Offspring differ from parents in complex ways.
These observations suggest moving from scalar \(x\) and \(w\) to vectors \(\mathbf{x} \in \mathbb{R}^n\) and \(\mathbf{w} \in \mathbb{R}^m\), and from covariance to a multilinear map. Tensors provide precisely this structure.
3. Tensorial Formulation of the Price Equation
3.1 Trait and Fitness Spaces
Let the population consist of \(N\) individuals. Each individual \(i\) has:
- A trait vector \(\mathbf{x}_i \in \mathcal{X} \subseteq \mathbb{R}^n\), where \(\mathcal{X}\) is the trait space.
- A fitness vector \(\mathbf{w}_i \in \mathcal{W} \subseteq \mathbb{R}^m\), where each component \(w_i^{(k)}\) represents a different aspect of fitness (e.g., viability, mating success, fertility).
- A total fitness scalar \(W_i = \sum_{k=1}^m w_i^{(k)}\), used for weighting averages.
The population mean trait vector is
3.2 The Selection Tensor
Contracting \(\mathcal{S}\) with \(\bar{\mathbf{x}}\) yields the selection-induced change:
3.3 The Transmission Tensor
Contracting \(\mathcal{T}\) with \(\bar{\mathbf{x}}\) gives the transmission-induced change:
3.4 The Tensorial Price Equation
This is a coordinate-free, multilinear generalization of the scalar Price equation.
3.5 Discrete vs. Continuous Formulations
In contrast, information geometric approaches typically require continuous distributions and smooth fitness landscapes. The continuous-time limit emerges under specific assumptions:
- Smooth trait distributions (e.g., Gaussian);
- Differentiable fitness functions;
- Small per-generation changes (\(|\Delta \mathbf{x}_i| \ll \sqrt{\operatorname{Var}(\mathbf{x})}\));
3.6 Metric Structure and Information Geometry
The trait space \(\mathcal{X}\) can be equipped with a metric tensor \(g_{ij}\), defining an inner product \(\langle \mathbf{u}, \mathbf{v} \rangle = g_{ij} u^i v^j\). This allows us to define the selection gradient as:
and relate it to the selection tensor via \(\mathcal{S} = \boldsymbol{\beta} \otimes \bar{\mathbf{x}}\) in the case of linear selection.
The Fisher information metric provides a natural choice for \(g_{ij}\):
4. A Unified Selection Model: Natural, Sexual, and Beyond
4.1 Natural Selection Tensor \(\mathcal{S}_N\)
Natural selection acts on viability and fertility. Let \(\mathbf{w}_N\) be the fitness component due to survival and reproduction. Then:
4.2 Sexual Selection Tensor \(\mathcal{S}_S\)
Sexual selection arises from mate choice or competition. Let \(\mathbf{w}_S\) be the fitness component due to mating success. We model this via a preference tensor \(\mathcal{P}\):
The sexual selection tensor becomes:
4.3 Subset Selection Tensor \(\mathcal{S}_B\)
Subset selection (e.g., choosing apples, books) is a special case where \(w_i' \in \{0,1\}\). Then:
4.4 Cultural Selection Tensor \(\mathcal{S}_C\)
Cultural selection (ideas, behaviors) may involve imitation bias, modeled via an imitation tensor \(\mathcal{I}\):
4.5 Composite Selection Tensor
When multiple selection processes act simultaneously, the total selection tensor is the sum:
4.6 Transmission Tensor \(\mathcal{T}\)
Transmission includes:
- Mutation: \(\Delta \mathbf{x}_i \sim \mathcal{N}(0, \mathcal{M})\), where \(\mathcal{M}\) is a mutation tensor.
- Recombination: modeled as a bilinear map \(\mathcal{R}(\mathbf{x}_{i1}, \mathbf{x}_{i2})\).
- Cultural transmission: modeled via a social learning tensor \(\mathcal{L}\).
Then:
5. Dynamical System for Multistage Selection
Price's Fig. 6a shows multistage selection with alternating property change and selection. In our framework, this is a discrete-time dynamical system:
If selection and transmission tensors are time-dependent (frequency-dependent, density-dependent, or environmentally driven), this becomes a nonlinear dynamical system. The continuous-time approximation is:
6. Integration with Kalman Filtering for Uncertain Dynamics
6.1 Kalman Filter Recap
The Kalman filter [5] provides optimal state estimation for linear dynamical systems with Gaussian noise. For a state vector \(\mathbf{z}_t\):
6.2 Selection-Kalman Model
The state vector comprises the mean trait vector and its covariance:
where \(\Sigma_t = \operatorname{Cov}(\mathbf{x}, \mathbf{x})\) is the trait covariance matrix.
The prediction step corresponds to redistribution (mutation, recombination): \(F_t = I + \mathcal{T}_t\).
The update step incorporates selection based on fitness observations: \(H_t = \nabla_{\bar{\mathbf{x}}} \ln \bar{w}(\bar{\mathbf{x}}_t, \Sigma_t)\).
7. Numerical Examples
7.1 Sexual Selection in 2D Trait Space
Consider two traits: body size (\(x_1\)) and ornamentation (\(x_2\)). Fitness has components:
Simulation shows rapid evolution of ornamentation due to sexual selection while body size stabilizes at the viability optimum.
7.2 Multistage Selection with Uncertainty
We simulate a population undergoing alternating selection and redistribution phases with environmental noise. The Selection-Kalman filter accurately tracks trait evolution while quantifying uncertainty.
8. Discussion
Price envisioned a "mathematical theory of selection" akin to Shannon's communication theory. Our tensor-based formalism moves toward that goal by providing a coordinate-free, multilinear language for selection, capable of describing everything from chemical selection to cultural evolution. The selection tensor \(\mathcal{S}\) and transmission tensor \(\mathcal{T}\) form the building blocks of a general selection calculus.
Limitations: While powerful, our framework has limitations: computational complexity (\(O(n^2)\) parameters), data requirements, and Gaussian assumptions in the Kalman filter. Future work could address these via sparse tensors, Bayesian methods, particle filters, and connections to evolutionary game theory.
9. Conclusion
We have developed a comprehensive tensorial generalization of the Price equation that provides a unified mathematical framework for selection processes. This framework generalizes Price's model to multiple traits, incorporates sexual and cultural selection, clarifies the selection-entropy relationship, provides dynamical systems for multistage selection, and integrates Kalman filtering for uncertainty quantification. Price wrote that "selection theory is a theory waiting to be born." With this tensorial formulation, we take a significant step toward its birth.
Appendix A. Mathematical Preliminaries on Tensors
Let \(V\) be an \(n\)-dimensional vector space over \(\mathbb{R}\) with basis \(\{\mathbf{e}_i\}\). A tensor of type \((p,q)\) is an element of \(V^{\otimes p} \otimes (V^*)^{\otimes q}\). In components, a tensor \(T\) is written as \(T_{j_1 \dots j_q}^{i_1 \dots i_p}\). Contraction is summing over one upper and one lower index, e.g., \(T_j^i v^j\). The outer product \(\mathbf{u} \otimes \mathbf{v}\) has components \(u^i v^j\).
Appendix B. Derivation of the Tensorial Price Equation
Starting from the scalar Price equation for each trait component \(x^j\):
and noting \(W = \sum_k w^{(k)}\), we have \(\operatorname{Cov}(W, x^j) = \sum_k \operatorname{Cov}(w^{(k)}, x^j)\). Thus:
Identifying \(S_k^j = \operatorname{Cov}(w^{(k)}, x^j) / \bar{W}\) and \(T^j = \mathbb{E}(W \Delta x^j) / \bar{W}\) gives the tensorial form.
Appendix C. Special Cases and Examples
C.1 Single trait, single fitness component: If \(n=m=1\), then \(\mathcal{S} = \operatorname{Cov}(w,x)/\bar{W}\) and \(\mathcal{T} = \mathbb{E}(w\Delta x)/\bar{W}\), reducing to the classical Price equation.
C.2 Linear selection gradient: Under \(W(\mathbf{x}) = \alpha + \boldsymbol{\beta}\cdot\mathbf{x}\), we have \(\mathcal{S} = \boldsymbol{\beta}\otimes\bar{\mathbf{x}}\).
C.3 Price's Fig. 5 example: For gene frequency \(x\), \(\mathcal{S}\) captures covariance between fecundity/survival and gene frequency, \(\mathcal{T}\) captures recombination/mutation.
Appendix D. Connection to Information Geometry
The Fisher information metric on a statistical manifold of trait distributions is
The selection gradient \(\boldsymbol{\beta}\) is the natural gradient of mean fitness. The selection-information inequality follows from the Cramér-Rao bound.
Appendix E. Selection-Kalman Algorithm Pseudocode
Algorithm: Selection-Kalman Filter
Input: Initial mean x̄₀, covariance Σ₀, noise Q, R
Output: Trajectory {x̄_t, Σ_t}
t ← 0
while t < T:
# Prediction (redistribution)
x̄_{t|t-1} ← (I + T_t) x̄_t
Σ_{t|t-1} ← (I + T_t) Σ_t (I + T_t)^T + Q_t
# Update (selection)
H_t ← ∇_{x̄} ln w̄(x̄_{t|t-1}, Σ_{t|t-1})
K_t ← Σ_{t|t-1} H_t^T (H_t Σ_{t|t-1} H_t^T + R_t)^{-1}
w_obs ← measurement + noise
x̄_{t+1} ← x̄_{t|t-1} + K_t (w_obs - w̄)
Σ_{t+1} ← (I - K_t H_t) Σ_{t|t-1}
t ← t+1
Appendix F. MATLAB Code for Selection-Kalman Filter
classdef selection_kalman < handle
properties
n_dim; x_mean; x_cov; price_tensor; Q; R;
history_mean; history_cov; generation;
end
methods
function obj = selection_kalman(n, x0, cov0, price, Q, R)
obj.n_dim = n; obj.x_mean = x0; obj.x_cov = cov0;
obj.price_tensor = price; obj.Q = Q; obj.R = R;
obj.history_mean = x0; obj.history_cov = cov0; obj.generation = 0;
end
function step(obj, redist_func, fitness_func)
% Prediction
[x_pred, J] = redist_func(obj.x_mean, obj.x_cov);
cov_pred = J * obj.x_cov * J' + obj.Q;
% Update
H = obj.fitness_gradient(x_pred, cov_pred, fitness_func);
S = H * cov_pred * H' + obj.R;
K = cov_pred * H' / S;
w_obs = fitness_func(x_pred, cov_pred) + mvnrnd(zeros(obj.n_dim,1), obj.R)';
obj.x_mean = x_pred + K * (w_obs - fitness_func(x_pred, cov_pred));
obj.x_cov = (eye(obj.n_dim) - K*H) * cov_pred;
% Store
obj.history_mean = [obj.history_mean; obj.x_mean'];
obj.history_cov = cat(3, obj.history_cov, obj.x_cov);
obj.generation = obj.generation + 1;
end
end
end
References
[1] Price, G. R. (1995). The nature of selection. Journal of Theoretical Biology, 175(3), 389–396.
[2] Price, G. R. (1970). Selection and covariance. Nature, 227(5257), 520–521.
[3] Price, G. R. (1972). Extension of covariance selection mathematics. Annals of Human Genetics, 35(4), 485–490.
[4] Shannon, C. E. (1948). A mathematical theory of communication. Bell System Technical Journal, 27(3), 379–423.
[5] Kalman, R. E. (1960). A new approach to linear filtering and prediction problems. Journal of Basic Engineering, 82, 35–45.
[6] Amari, S. (1985). Differential-geometrical methods in statistics. Springer.
[7] Amari, S., & Nagaoka, H. (2000). Methods of information geometry. American Mathematical Society.
[8] Frank, S. A. (2009). Natural selection maximizes Fisher information. Journal of Evolutionary Biology, 22(2), 231–244.
[9] Lande, R. (1979). Quantitative genetic analysis of multivariate evolution. Evolution, 33, 402–416.
© 2025 Sourav Kundu et al. All rights reserved.