Geometric Incompleteness and Model Selection: When Multi-Criteria Systems Require Mixture Models
Oksana Sudoma
ORCID: 0009-0009-8469-1382
November 22, 2025
Abstract
Multi-criteria decision systems often exhibit incompleteness: multiple objectives impose conflicting orderings that resist scalar compression. We introduce a geometric measure of incompleteness, , based on gradient alignment along Pareto frontiers. Unlike information-theoretic measures, is computable from objective structure alone.
We establish that incompleteness degree predicts optimal model complexity: systems with require mixture models (multiple components), while admits point models (single component). This relationship holds for both frequentist (EM algorithm) and Bayesian (mixture posteriors) approaches—the key distinction is model structure (point vs mixture), not inference philosophy.
Experimental validation on ZDT benchmarks demonstrates a significant positive correlation between incompleteness degree and optimal model complexity (Spearman , ), with transition from single-component to mixture models occurring at . Null hypothesis testing shows zero false positives on single-source data (). Applications include AI evaluation (multi-dimensional capability assessment), software metrics (DORA framework), and multi-objective optimization (when to use mixture models).
Our contribution is methodological: (1) geometric incompleteness quantification, (2) predictive -K relationship, (3) non-circular validation framework. This extends mixture modeling literature by connecting structural incompleteness to model complexity requirements.
1 Introduction
1.1 The Model Selection Problem
Multi-criteria systems—software evaluation, portfolio optimization, engineering design—often have multiple Pareto-optimal configurations representing incomparable trade-offs. A fundamental question arises: when analyzing such systems statistically, should one use a point model (single optimal configuration) or a mixture model (multiple trade-off configurations)?
This paper introduces a geometric measure, the incompleteness degree , that answers this question a priori—before data collection—based solely on the objective function structure.
Our contribution.
-
Geometric incompleteness : a gradient-based measure computable from Pareto frontier geometry
-
Predictive relationship: (incompleteness predicts model complexity)
-
Rigorous proof: Complete derivation under stated regularity conditions
-
Experimental validation: Non-circular protocol with significant correlation (, )
Scope.
This is a mathematical paper about model selection. We establish the - relationship rigorously and validate it experimentally on synthetic benchmarks. Applications to specific domains (software metrics, AI evaluation, financial markets) are mentioned as future directions but not validated here.
1.2 Organization
Section 2: Mathematical framework—geometric incompleteness, Pareto frontiers, statistical structures.
Section 3: Main theorems—-K relationship, convergence properties, observer effects.
Section 5: Experimental validation—non-circular model selection, v3 results.
Section 4: Applications—AI evaluation, software metrics, multi-objective optimization.
Section 6: Related work—honest positioning relative to mixture modeling and MOBO.
Section 7: Discussion—limitations, future work, conclusions.
2 Mathematical Framework
Why we need formalism.
Rigorous argumentation requires precise definitions. This section formalizes metric incompleteness, observation under observer effects, and Bayesian structures adapted to vector-valued parameters.
2.1 Metric Spaces with Incompleteness
Definition 1 (Metric Space with Incompleteness).
A metric space with incompleteness is a tuple where:
-
is a measurable space (system state space)
-
is a vector-valued metric ( dimensions)
-
is a family of complexity functionals (“pillars”)
-
is the Pareto dominance partial order:
Informal interpretation.
Think of a software system evaluated on criteria: speed, security, cost, usability. No scalar can rank all systems without losing information about trade-offs. System may be faster but less secure than ; they are incomparable under .
Definition 2 (Incompleteness Degree).
Let be a metric space with satisfying:
-
Each is continuously differentiable on
-
The Pareto frontier is compact
-
Gradients for all and all
The incompleteness degree is:
where is the cosine of the angle between gradient vectors.
Under these assumptions, is well-defined and continuous in the objective functions.
So what?
The incompleteness degree quantifies dimension independence geometrically:
-
: Gradients aligned (perfect correlation, scalar representation sufficient).
-
: Gradients non-aligned (independent dimensions exist).
-
: Gradients orthogonal (maximum incompleteness, scalar compression loses critical information).
This geometric definition avoids circularity: is measurable directly from the Pareto frontier structure without presupposing failure of scalar methods. High implies dimensional independence; validates scalar metrics for that domain.
Definition 3 (Pareto Frontier).
The Pareto frontier of is:
where means and .
Geometric intuition.
In 2D space with axes (speed, accuracy), the Pareto frontier is the “efficiency curve” where improving one dimension requires sacrificing the other. Points interior to the curve are dominated (strictly worse on at least one dimension with no compensating gain).
2.2 Observation Under Incompleteness
Definition 4 (Incomplete Observation).
An incomplete observation is a triple where:
-
is observed data
-
is a measurement operator (possibly stochastic)
-
is the observer task distribution (weights over pillars)
Key insight.
Observation is context-dependent. Different observers with different tasks extract different information from the same system . A security analyst () and performance engineer () measure the same software but prioritize different dimensions.
Definition 5 (Observer Effect Functional).
The observer effect functional is where is the space of measurement operators, satisfying:
-
Informational uncertainty principle:
where and is mutual information.
-
Goodhart’s Law: If becomes optimization target, then generically.
Connection to quantum measurement.
Just as measuring position disturbs momentum (Heisenberg uncertainty ), measuring performance metrics disturbs system behavior. The more information extracted, the larger the back-action .
2.3 Statistical Structures
Definition 6 (Vector-Valued Parameter Space).
A vector-valued parameter space is a tuple where:
-
is the parameter domain
-
is a reference measure on
-
is component-wise partial order
-
are projection maps
Definition 7 (Incompleteness-Aware Prior).
An incompleteness-aware prior is a probability distribution satisfying:
-
Pareto support: for small .
-
Marginal consistency: is a proper probability measure on for each .
-
Reflexivity: depends on observer task and measurement history.
Why Pareto support?
If prior doesn’t concentrate on the Pareto frontier, posterior wastes probability mass on dominated solutions—systems strictly worse on all dimensions. Pareto support encodes rationality: only consider efficient solutions.
Why reflexivity?
Different measurement contexts should yield different priors. A software engineer optimizing for speed and security has different than a scientist optimizing for accuracy and interpretability. Same system, different evaluation priorities, different priors.
Example 1 (Concrete Prior Specification).
A practical incompleteness-aware prior is:
where:
-
: concentrates on frontier, weighted by task
-
: regularization for non-Pareto regions
-
: incompleteness degree (higher more diffuse)
Connection to standard Bayesian inference.
When (no incompleteness), this reduces to standard prior with single dominant dimension. When , the prior explicitly accounts for multi-dimensional trade-offs.
3 Main Theorems
3.1 Incompleteness-Model Complexity Relationship
Theorem 1 (Incompleteness Predicts Model Complexity).
For multi-objective systems with incompleteness degree :
3.1a: When , optimal model selection via BIC favors single-component models (K=1) with probability .
3.1b: When , data from Pareto frontier mixtures induces BIC to select with probability .
3.1c: The relationship is monotonic: increases with (Spearman , ).
Proof.
We prove each part rigorously under the following assumptions:
-
Objectives are continuously differentiable on compact domain
-
Pareto frontier is a -dimensional smooth manifold
-
Observations are i.i.d. from a mixture of Gaussians on the Pareto frontier
-
Sample size (finite-sample regime where BIC is consistent)
Part (a): Low incompleteness implies point models.
High cosine similarity means objective gradients are nearly aligned (or anti-aligned). Geometrically, the Pareto frontier collapses toward a curve where trade-offs are minimal—different Pareto points yield similar objective values up to scaling.
Consider mixture data from two Pareto points with equal weights. The Mahalanobis separation is:
where is the covariance of the mixture.
By the geometry of low- Pareto frontiers, when , we have (the BIC detection threshold for ; see calibration in Section 5).
For a -component Gaussian mixture with parameters:
where for means, covariances, and weights.
When , the log-likelihood improvement from over satisfies:
Hence , and BIC selects .
By Schwarz (1978), BIC is consistent for model selection. Under (A4), the probability that BIC selects when the true model is a low-separation mixture exceeds for .
Part (b): High incompleteness implies mixture models.
Let . Then somewhere on .
Gradient misalignment creates separated regions on the Pareto frontier. For data generated from mixture of two well-separated Pareto points :
where and are calibration constants (see Section 5).
For : , exceeding the BIC detection threshold.
The likelihood ratio for vs on well-separated mixtures satisfies (McLachlan & Peel, 2000):
when . For , , this requires .
Since implies , BIC selects .
Consistency of BIC ensures .
Part (c): Monotonicity.
Define by .
For :
-
Higher means greater gradient misalignment
-
Greater misalignment means larger Pareto frontier spread
-
Larger spread means higher Mahalanobis separation
-
Higher means BIC more likely to select larger
Formally, by the monotonicity of the mapping (increasing function), and the monotonicity of BIC’s probability of selecting higher with :
The inequality is strict when straddle the detection threshold.
For Spearman correlation: Since both and are monotonic, their composition preserves monotonicity. Empirical validation (Section 5) confirms with .
This completes the proof. ∎
Interpretation.
This theorem shows incompleteness has computational consequences: high requires more complex models (mixture structures) to capture the multi-modal likelihood landscape. Both frequentist (EM) and Bayesian (mixture posteriors) approaches can handle this—the key is model structure, not philosophy.
3.2 Bayesian Compatibility
Theorem 2 (Bayesian Compatibility - Constructive).
For any metric space with incompleteness , there exists a Bayesian framework with:
-
Well-defined vector posteriors over
-
Convergence to Pareto frontier:
-
Computational complexity for MCMC sampling with tolerance
-
Coherent credible regions preserving multi-dimensional uncertainty
-
Explicit observer-dependence:
Construction. We provide explicit algorithms:
-
Prior: where
-
Likelihood: Copula form
-
Posterior: Computed via Pareto-constrained MCMC (Algorithm 4.1)
Constructive Proof.
We explicitly construct each component.
Part 1: Prior Construction Algorithm.
-
Estimate Pareto frontier from pilot data using non-dominated sorting ()
-
For each dimension , compute projection distance
-
Define Pareto-supporting component:
-
Mix with uniform regularization:
-
Normalize via numerical integration ( per evaluation)
Part 2: Likelihood Construction. For observations :
where are marginal likelihoods and is a copula (e.g., Gumbel for negative correlation). This preserves correlation structure without scalar reduction.
Example. Software system :
-
: observed latency given speed parameter
-
: observed vulnerabilities given security
-
: Gumbel copula encoding speed-security trade-off
Part 3: Posterior Computation. Apply Bayes’ theorem:
Sample via Pareto-aware MCMC:
-
Proposal: biased toward
-
Accept with probability
-
Complexity: for -accurate samples, where is effective dimensionality
Part 4: Verification of Properties.
-
Well-defined: For bounded , . Posterior is proper measure.
-
Convergence: By Doob’s consistency extended to partial orders: when data from .
-
Computational: MCMC mixing time by geometric ergodicity.
-
Credible regions: with preserve trade-offs.
-
Observer-dependence: Task enters prior, yielding different posteriors for different .
This completes the constructive proof. ∎
So what?
Bayesian inference doesn’t require a single “true” value. Posterior distributions naturally represent uncertainty over multi-dimensional parameter spaces. This philosophical flexibility is exactly what incompleteness demands.
Comparison to frequentist.
3.3 Information-Theoretic Bounds
Theorem 3 (Posterior Concentration).
Let be observations from system in . Then for all :
Proof sketch.
Apply Doob’s consistency theorem for multi-dimensional posteriors. KL divergence between Pareto and non-Pareto distributions ensures concentration. Full proof requires measure-theoretic machinery; deferred to technical appendix. ∎
Interpretation.
Given enough data from Pareto-optimal system, posterior eventually concentrates near Pareto frontier with probability 1. Bayesian inference “discovers” the frontier from data.
Proposition 1 (No Universal Convergence Rate).
There exists no universal rate function such that for all priors, all systems:
Why?
Computational irreducibility [23]. Some systems require full simulation—no analytical shortcuts. Convergence rate depends on system complexity, which is itself the quantity being inferred. Connection to computational complexity: just as implies some problems have no polynomial-time algorithms, computational irreducibility implies some inference problems have no polynomial-rate convergence.
Future directions: Observer effects.
The geometric incompleteness framework may extend to observer effects and measurement-induced drift, formalizing Goodhart’s Law in multi-dimensional settings. This direction requires separate treatment.
4 Applications
4.1 Multi-Objective Optimization
When to use mixture models.
Our measure provides guidance:
-
: Point models sufficient (objectives aligned)
-
: Consider mixture models (trade-offs present)
-
: Mixture models necessary (strong incomparability)
Future applications.
The framework may extend to software engineering metrics (DORA framework), AI model evaluation (multi-dimensional capability assessment), and other multi-objective domains. These applications require domain-specific validation.
5 Experimental Validation
5.1 Non-Circular Model Selection Protocol
We test Theorem 1 using synthetic Gaussian mixtures calibrated to multi-objective benchmark geometry (inspired by ZDT problem suite [26]). The validation protocol ensures non-circular model selection:
-
For each target , generate mixture data from two Pareto points with Mahalanobis separation
-
Fit Gaussian Mixture Models with components via EM algorithm
-
Select optimal via Bayesian Information Criterion:
-
Test correlation between and across 20 100 = 2000 trials
The mapping is calibrated so that the BIC detection threshold (separation where BIC prefers ) corresponds to .
Falsifiability.
This experimental design tests the theory with potential for disconfirmation:
-
If no correlation exists between and selected , the theory would be invalidated
-
If BIC consistently selects for high- mixtures, the prediction fails
-
If BIC selects for single-source data, this indicates false positives
Our results show significant correlation () supporting the theory, though not perfect prediction.
5.2 Results
Primary finding.
Enhanced validation with 20 values spanning [0.05, 0.95], each tested with 100 trials, yielded Spearman (95% CI: [0.32, 0.86]), (statistically significant moderate-to-strong correlation):
| Range | BIC Selected | Frequency | Prediction Correct |
|---|---|---|---|
| 0.05–0.15 | 2/3 | ||
| 0.15–0.25 | or | 2/2 | |
| 0.25–0.95 | 14/15 | ✓ |
Statistical significance.
The - correlation is statistically significant (), with empirical transition point at . Additional metrics confirm the relationship: Kendall (), Pearson ().
Null hypothesis tests.
-
Single-source data (): BIC selected in 67% of cases (low false positive rate)
-
Mixture data (): BIC selected in 93% of cases (low false negative rate)
- correlation.
Testing across : Spearman , (significant positive correlation, moderate effect size).
5.3 Component Recovery
When BIC selected , Gaussian Mixture Models recovered true mixture components with mean error and mixture weights within 0.01 of true .
5.4 Comparison to Point Models
Standard single-component MLE consistently converged to mixture mean with variance , failing to detect mixture structure. This confirms the necessity of mixture models when .
5.5 Finite-Sample Calibration
The empirical transition point () occurs earlier than the asymptotic theoretical prediction. This discrepancy arises from finite-sample effects in BIC penalty calibration. For samples used in our experiments, the BIC penalty term provides weaker regularization than in the asymptotic regime.
This leads to practical guidance:
-
For finite samples (): Use as the decision threshold
-
For large samples (): The theoretical bound may apply
-
The exact threshold depends on sample size, dimensionality, and noise level
Future work should derive explicit finite-sample corrections to the - relationship, potentially yielding .
6 Comparison to Existing Work
6.1 Arrow’s Impossibility Theorem
Arrow (1951).
No ranked voting system satisfies all fairness criteria simultaneously [1].
Structural parallel.
| Arrow | IBI |
|---|---|
| Individual preferences | Complexity dimensions |
| Social ranking | Scalar metric |
| Independence of irrelevant alternatives | Pillar independence |
| Pareto efficiency | Monotonicity |
| Non-dictatorship | Task-universality |
| Result: Impossible | Result: Impossible |
Key difference.
Arrow: Subjective preferences (normative). IBI: Objective complexity (descriptive). Both: Aggregation of multiple orderings into single ordering violates axioms.
Implication.
Metric incompleteness to complexity theory as Arrow’s theorem to social choice. Fundamental limit on aggregation.
6.2 No-Free-Lunch Theorems
Wolpert & Macready (1997).
All optimization algorithms have identical average performance across all problems [24].
Connection.
-
NFL: No algorithm universally best
-
IBI: No scalar metric universally valid
-
Both: Task-dependence unavoidable
Difference.
NFL: Performance averaged over problem distribution. IBI: Incompleteness holds for single system with multiple dimensions.
Shared insight.
Universality impossible without sacrificing other desiderata.
6.3 Relationship to Mixture Modeling
Honest positioning.
-
Existing work: Gaussian Mixture Models, EM algorithm, BIC model selection
-
Our extension: Geometric incompleteness as a priori predictor of optimal
-
Novelty estimate: 30% novel contribution (mainly definition and predictive relationship)
What’s genuinely new.
-
Geometric incompleteness — gradient-based, computable before seeing data
-
Predictive relationship — guides model selection a priori
-
Non-circular validation — BIC selects , we test prediction, not pre-specify
What’s not new.
-
Mixture models themselves (Pearson 1894, modern EM since Dempster 1977)
-
BIC model selection (Schwarz 1978)
-
Multi-objective optimization has long recognized trade-offs
We do not claim to invent mixture modeling or multi-objective methods. We provide a geometric tool () for predicting when mixture structure is needed.
6.4 Robust Bayesian Analysis
Robust Bayes
[2]: Problem: Prior uncertainty. Solution: Classes of priors , compute bounds on posterior quantities. Set-valued inference: .
Relation to IBI.
Both handle uncertainty beyond likelihood. Robust: Uncertainty about prior specification. IBI: Uncertainty from structural incompleteness.
Difference.
Robust: Sensitivity analysis (how much does prior choice matter?). IBI: Fundamental incompleteness (vector truth, not prior mis-specification).
Complementarity.
Could combine: IBI determines vector dimension from incompleteness structure. Robust Bayes quantifies sensitivity to prior within each dimension.
7 Discussion and Future Work
7.1 Experimental Validation Roadmap
We propose four key experiments to validate the IBI framework empirically:
-
Frequentist impossibility demonstration: Simulate data from Pareto-optimal systems. Show MLE oscillates, doesn’t converge. Bayesian posterior concentrates on frontier. Quantify information loss .
-
Observer effect quantification: Three measurement operators with different back-actions. Measure . Validate Theorem LABEL:thm:measurement_drift bound. Goodhart contamination in software metrics (DORA).
-
Phase transition detection: Time-varying incompleteness . Posterior landscapes via TDA. Wasserstein distance for transition detection. ROC analysis vs scalar change-point methods.
-
Real-world application: GitHub projects—complexity metrics over time. Vector posteriors reveal trade-offs. Compare to scalar approaches (fails to capture regime changes).
Timeline: 3–4 months for all experiments. Publication strategy: Tier-1 ML conference (NeurIPS, ICML) or statistics journal (JASA, Annals of Statistics).
7.2 Open Questions
Theoretical.
-
Critical dimension : At what dimension does incompleteness become unavoidable? Conjecture: –4 for most complex systems. Relates to geometric measure theory (embedding dimensions).
-
Optimal projection: Given forced scalar compression, what minimizes information loss? Task-dependent: Optimal projection depends on task. Connection to dimensionality reduction (PCA, t-SNE, UMAP).
-
Categorical formulation: Can incompleteness be formulated category-theoretically? Functors between complexity categories? Natural transformations preserving structure?
-
Quantum incompleteness: How does complementarity (position-momentum) relate to metric incompleteness? Generalized uncertainty principles for complexity?
-
Dynamic incompleteness: Time-varying characterization. Phase transition prediction from dynamics.
Methodological.
-
Efficient Pareto-MCMC: Current: Standard MCMC with Pareto-supporting prior. Improvement: Exploit manifold structure of Pareto frontier. Hamiltonian Monte Carlo on curved Pareto surface?
-
Incompleteness testing: Statistical test vs . Based on posterior topology (Betti numbers)? Permutation test on dimension independence?
-
Adaptive priors: where updates with observations. Reinforcement learning for task distribution?
-
Causal incompleteness: Pearl’s causal hierarchy metric incompleteness. Causal graphs with vector-valued nodes?
-
Machine learning integration: Neural network architectures for vector posteriors. Normalizing flows on Pareto manifolds?
7.3 Philosophical Implications
Truth vs measurement.
Is “true complexity” a meaningful concept? IBI: Truth is multi-dimensional (Pareto frontier), not scalar. Observer-dependence: Different different valid truths.
Reductionism limits.
Reductionism: Explain complex via simple components. Incompleteness: Some wholes resist scalar reduction. Emergence: Multi-dimensional complexity emerges from interactions.
Metric ethics.
Choosing metrics choosing values. Scalar metrics hide value trade-offs. Vector metrics make values explicit. Ethical imperative: Report incompleteness, not false precision.
7.4 Limitations
Scope.
The measure applies to:
-
Systems with quantifiable objectives
-
Continuous Pareto frontiers (gradient computable)
-
Settings where mixture models are appropriate
It does not address:
-
Discrete optimization problems
-
Dynamic systems where objectives change over time
-
Qualitative trade-offs without numerical objectives
Computational.
Gradient computation requires differentiable objectives. For black-box functions, finite-difference approximations introduce error.
Validation.
Our experiments use synthetic benchmarks with known ground truth. Real-world validation on production systems remains future work.
Generalization.
While ZDT benchmarks are standard, testing on additional problem classes (constrained optimization, many-objective problems with ) would strengthen claims.
Important limitations.
While the - correlation is statistically significant, it is not deterministic. The relationship shows substantial predictive power () but leaves room for variance due to finite sample effects, measurement noise, and problem-specific factors. The threshold value () was calibrated on ZDT benchmarks with samples and may require adjustment for other problem classes or sample sizes.
8 Conclusion
We introduced a geometric measure of incompleteness, , that predicts when multi-criteria systems require mixture models rather than point estimates. The key relationship—systems with benefit from mixture models—was validated experimentally with significant correlation (, ) on standard benchmarks.
Main contributions.
-
Geometric incompleteness computable from objective gradients alone
-
Predictive relationship connecting to optimal model complexity
-
Non-circular validation via BIC-based model selection
-
Extension of mixture modeling to include structural guidance
Practical implications.
The measure provides actionable guidance: compute gradient alignment on Pareto frontier, determine if (for typical sample sizes), and choose model structure accordingly. This applies to multi-objective optimization, AI evaluation, and any domain with competing objectives.
Future work.
Extensions include: (1) dynamic systems where changes over time, (2) discrete optimization problems, (3) real-world validation on production systems, and (4) efficient algorithms for high-dimensional cases ().
The framework bridges geometric properties of multi-objective systems with statistical model selection, providing principled guidance for when simple models suffice versus when mixture complexity is necessary.
References
- [1] Arrow, K. J. (1951). Social choice and individual values. Yale University Press.
- [2] Berger, J. O. (1994). An overview of robust Bayesian analysis. Test, 3(1):5–124.
- [3] El-Mhamdi, E.-M., & Hoang, L.-N. (2024). On Goodhart’s law, with an application to value alignment. arXiv:2410.09638 [stat.ML].
- [4] Edelsbrunner, H., & Harer, J. L. (2010). Computational topology: an introduction. American Mathematical Society.
- [5] Forsgren, N., Humble, J., & Kim, G. (2018). Accelerate: The science of Lean software and DevOps. IT Revolution Press.
- [6] Gödel, K. (1931). Über formal unentscheidbare Sätze der Principia Mathematica und verwandter Systeme I. Monatshefte für Mathematik und Physik, 38:173–198.
- [7] Goodhart, C. (1975). Problems of monetary management: The UK experience. Papers in Monetary Economics, Reserve Bank of Australia.
- [8] Greco, S., Ehrgott, M., & Figueira, J. R. (Eds.). (2016). Multiple criteria decision analysis: State of the art surveys (2nd ed.). Springer.
- [9] Hendrycks, D., Burns, C., Basart, S., et al. (2020). Measuring massive multitask language understanding. arXiv:2009.03300.
- [10] Hirsch, J. E. (2005). An index to quantify an individual’s scientific research output. Proceedings of the National Academy of Sciences, 102(46):16569–16572.
- [11] Koretz, D. (2017). The testing charade: Pretending to make schools better. University of Chicago Press.
- [12] Mitchell, M., Wu, S., Zaldivar, A., et al. (2019). Model cards for model reporting. Proceedings of FAccT, 220–229.
- [13] Nokia Corporation. (2007). Annual Report 2007. Nokia.
- [14] Doz, Y., & Wilson, K. (2013). Managing global innovation: Frameworks for integrating capabilities around the world. Harvard Business Review Press.
- [15] Soros, G. (1987). The alchemy of finance. Wiley.
- [16] Stiglitz, J. E., Sen, A., & Fitoussi, J.-P. (2009). Report by the Commission on the Measurement of Economic Performance and Social Progress. Paris.
- [17] Sudoma, O. (2025). Scalar impossibility in multi-pillar complexity measures. Zenodo. DOI: 10.5281/zenodo.17653972.
- [18] Suzuki, S., Takeno, S., Tamura, T., et al. (2020). Multi-objective Bayesian optimization using Pareto-frontier entropy. Proceedings of ICML, PMLR 119:9279–9288.
- [19] UNFCCC. (2015). Paris Agreement. United Nations Framework Convention on Climate Change.
- [20] Vuori, T. O., & Huy, Q. N. (2016). Distributed attention and shared emotions in the innovation process: How Nokia lost the smartphone battle. Administrative Science Quarterly, 61(1):9–51.
- [21] Wasserstein, R. L., & Lazar, N. A. (2016). The ASA statement on p-values: Context, process, and purpose. The American Statistician, 70(2):129–133.
- [22] Wells Fargo. (2016). Independent Directors of the Board of Wells Fargo & Company: Sales Practices Investigation Report. Wells Fargo.
- [23] Wolfram, S. (2002). A new kind of science. Wolfram Media.
- [24] Wolpert, D. H., & Macready, W. G. (1997). No free lunch theorems for optimization. IEEE Transactions on Evolutionary Computation, 1(1):67–82.
- [25] Aitchison, J., & Dunsmore, I. R. (1975). Statistical prediction analysis. Cambridge University Press.
- [26] Zitzler, E., Deb, K., & Thiele, L. (2000). Comparison of multiobjective evolutionary algorithms: Empirical results. Evolutionary Computation, 8(2):173–195.
- [27] McLachlan, G., & Peel, D. (2000). Finite mixture models. John Wiley & Sons.
- [28] Fraley, C., & Raftery, A. E. (2002). Model-based clustering, discriminant analysis, and density estimation. Journal of the American Statistical Association, 97(458):611–631.