TY - JOUR
T1 - Generalization power of threshold Boolean networks
AU - Ruz, Gonzalo A.
AU - Cho, Anthony D.
N1 - Publisher Copyright:
© 2025 Elsevier B.V.
PY - 2025/11
Y1 - 2025/11
N2 - Threshold Boolean networks are widely used to model gene regulatory systems and social dynamics such as consensus formation. In these networks, each node takes a binary value (0 or 1), leading to an exponential growth in the number of possible configurations with the number of nodes (2n). Inferring such networks involves learning a weight matrix and threshold vector from configuration data. However, in practice, the full state transition matrix is rarely available. This study investigates the generalization power of threshold Boolean networks, specifically, their ability to accurately infer the underlying network as the amount of available training data is reduced or degraded. We conducted experiments to empirically explore this generalization across networks with varying sizes and connectivities, using the perceptron learning algorithm for training. We also examined scenarios where data is degraded and evaluated the networks’ ability to preserve the original system's fixed points. Our results reveal an inverse relationship between network size and the required portion of the state transition matrix: larger networks require less data to infer the original structure. For example, networks with five nodes required about 62.5% of the data, whereas networks with nine nodes needed only 46%. Conversely, we observed a positive correlation between node indegree and the amount of training data necessary for accurate inference. In terms of preserving fixed points, our findings indicate that using approximately 40% of the data is generally sufficient to retain the fixed points present in the complete dataset.
AB - Threshold Boolean networks are widely used to model gene regulatory systems and social dynamics such as consensus formation. In these networks, each node takes a binary value (0 or 1), leading to an exponential growth in the number of possible configurations with the number of nodes (2n). Inferring such networks involves learning a weight matrix and threshold vector from configuration data. However, in practice, the full state transition matrix is rarely available. This study investigates the generalization power of threshold Boolean networks, specifically, their ability to accurately infer the underlying network as the amount of available training data is reduced or degraded. We conducted experiments to empirically explore this generalization across networks with varying sizes and connectivities, using the perceptron learning algorithm for training. We also examined scenarios where data is degraded and evaluated the networks’ ability to preserve the original system's fixed points. Our results reveal an inverse relationship between network size and the required portion of the state transition matrix: larger networks require less data to infer the original structure. For example, networks with five nodes required about 62.5% of the data, whereas networks with nine nodes needed only 46%. Conversely, we observed a positive correlation between node indegree and the amount of training data necessary for accurate inference. In terms of preserving fixed points, our findings indicate that using approximately 40% of the data is generally sufficient to retain the fixed points present in the complete dataset.
KW - Discrete dynamical system
KW - Gene regulatory networks
KW - Generalization power
KW - Perceptron
KW - Threshold Boolean networks
UR - https://www.scopus.com/pages/publications/105015476358
U2 - 10.1016/j.biosystems.2025.105572
DO - 10.1016/j.biosystems.2025.105572
M3 - Article
C2 - 40885514
AN - SCOPUS:105015476358
SN - 0303-2647
VL - 257
JO - BioSystems
JF - BioSystems
M1 - 105572
ER -