Version 1
: Received: 14 February 2022 / Approved: 21 February 2022 / Online: 21 February 2022 (14:13:26 CET)
Version 2
: Received: 7 June 2023 / Approved: 7 June 2023 / Online: 7 June 2023 (13:20:18 CEST)
Version 3
: Received: 19 September 2024 / Approved: 19 September 2024 / Online: 20 September 2024 (09:42:10 CEST)
How to cite:
Gogoshin, G.; Rodin, A. Minimum Statistical Uncertainty as Bayesian Network Model Selection Principle. Preprints2022, 2022020254. https://doi.org/10.20944/preprints202202.0254.v3
Gogoshin, G.; Rodin, A. Minimum Statistical Uncertainty as Bayesian Network Model Selection Principle. Preprints 2022, 2022020254. https://doi.org/10.20944/preprints202202.0254.v3
Gogoshin, G.; Rodin, A. Minimum Statistical Uncertainty as Bayesian Network Model Selection Principle. Preprints2022, 2022020254. https://doi.org/10.20944/preprints202202.0254.v3
APA Style
Gogoshin, G., & Rodin, A. (2024). Minimum Statistical Uncertainty as Bayesian Network Model Selection Principle. Preprints. https://doi.org/10.20944/preprints202202.0254.v3
Chicago/Turabian Style
Gogoshin, G. and Andrei Rodin. 2024 "Minimum Statistical Uncertainty as Bayesian Network Model Selection Principle" Preprints. https://doi.org/10.20944/preprints202202.0254.v3
Abstract
Background: Bayesian Network (BN) modeling is a prominent methodology in computational systems biology. However, the incommensurability of datasets fre- quently encountered in life science domains gives rise to contextual dependence and numerical irregularities in the behavior of model selection criteria (such as MDL, Minimum Description Length) used in BN reconstruction. This renders model features, first and foremost dependency strengths, incomparable and diffi- cult to interpret. In this study, we derive and evaluate a model selection principle that addresses these problems. Results: The objective of the study is attained by (i) approaching model eval- uation as a classification problem, (ii) estimating the effect that sampling error has on the satisfiability of conditional independence criterion, as reflected by Mutual Information, and (iii) utilizing this error estimate to penalize uncertainty with the novel Minimum Uncertainty (MU) model selection principle. We vali- date our findings numerically and demonstrate the performance advantages of the MU criterion. Finally, we illustrate the advantages of the new model evaluation framework on real data examples. Conclusions: The new BN model selection principle successfully overcomes per- formance irregularities observed with MDL, offers a superior average convergence rate in BN reconstruction, and improves the interpretability and universality of resulting BNs, thus enabling direct inter-BN comparisons and evaluations.
Computer Science and Mathematics, Applied Mathematics
Copyright:
This is an open access article distributed under the Creative Commons Attribution License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.